AdvSimd.Arm64 Class
Definition
Important
Some information relates to prerelease product that may be substantially modified before it’s released. Microsoft makes no warranties, express or implied, with respect to the information provided here.
Provides access to the ARM AdvSIMD hardware instructions, that are only available to 64-bit processes, via intrinsics.
public: ref class AdvSimd::Arm64 abstract : System::Runtime::Intrinsics::Arm::ArmBase::Arm64
public abstract class AdvSimd.Arm64 : System.Runtime.Intrinsics.Arm.ArmBase.Arm64
type AdvSimd.Arm64 = class
inherit ArmBase.Arm64
Public MustInherit Class AdvSimd.Arm64
Inherits ArmBase.Arm64
- Inheritance
- Derived
Properties
IsSupported |
Methods
Abs(Vector128<Double>) |
float64x2_t vabsq_f64 (float64x2_t a) A64: FABS Vd.2D, Vn.2D |
Abs(Vector128<Int64>) |
int64x2_t vabsq_s64 (int64x2_t a) A64: ABS Vd.2D, Vn.2D |
AbsoluteCompareGreaterThan(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcagtq_f64 (float64x2_t a, float64x2_t b) A64: FACGT Vd.2D, Vn.2D, Vm.2D |
AbsoluteCompareGreaterThanOrEqual(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcageq_f64 (float64x2_t a, float64x2_t b) A64: FACGE Vd.2D, Vn.2D, Vm.2D |
AbsoluteCompareGreaterThanOrEqualScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcage_f64 (float64x1_t a, float64x1_t b) A64: FACGE Dd, Dn, Dm |
AbsoluteCompareGreaterThanOrEqualScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcages_f32 (float32_t a, float32_t b) A64: FACGE Sd, Sn, Sm |
AbsoluteCompareGreaterThanScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcagt_f64 (float64x1_t a, float64x1_t b) A64: FACGT Dd, Dn, Dm |
AbsoluteCompareGreaterThanScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcagts_f32 (float32_t a, float32_t b) A64: FACGT Sd, Sn, Sm |
AbsoluteCompareLessThan(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcaltq_f64 (float64x2_t a, float64x2_t b) A64: FACGT Vd.2D, Vn.2D, Vm.2D |
AbsoluteCompareLessThanOrEqual(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcaleq_f64 (float64x2_t a, float64x2_t b) A64: FACGE Vd.2D, Vn.2D, Vm.2D |
AbsoluteCompareLessThanOrEqualScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcale_f64 (float64x1_t a, float64x1_t b) A64: FACGE Dd, Dn, Dm |
AbsoluteCompareLessThanOrEqualScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcales_f32 (float32_t a, float32_t b) A64: FACGE Sd, Sn, Sm |
AbsoluteCompareLessThanScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcalt_f64 (float64x1_t a, float64x1_t b) A64: FACGT Dd, Dn, Dm |
AbsoluteCompareLessThanScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcalts_f32 (float32_t a, float32_t b) A64: FACGT Sd, Sn, Sm |
AbsoluteDifference(Vector128<Double>, Vector128<Double>) |
float64x2_t vabdq_f64 (float64x2_t a, float64x2_t b) A64: FABD Vd.2D, Vn.2D, Vm.2D |
AbsoluteDifferenceScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vabd_f64 (float64x1_t a, float64x1_t b) A64: FABD Dd, Dn, Dm |
AbsoluteDifferenceScalar(Vector64<Single>, Vector64<Single>) |
float32_t vabds_f32 (float32_t a, float32_t b) A64: FABD Sd, Sn, Sm |
AbsSaturate(Vector128<Int64>) |
int64x2_t vqabsq_s64 (int64x2_t a) A64: SQABS Vd.2D, Vn.2D |
AbsSaturateScalar(Vector64<Int16>) |
int16_t vqabsh_s16 (int16_t a) A64: SQABS Hd, Hn |
AbsSaturateScalar(Vector64<Int32>) |
int32_t vqabss_s32 (int32_t a) A64: SQABS Sd, Sn |
AbsSaturateScalar(Vector64<Int64>) |
int64_t vqabsd_s64 (int64_t a) A64: SQABS Dd, Dn |
AbsSaturateScalar(Vector64<SByte>) |
int8_t vqabsb_s8 (int8_t a) A64: SQABS Bd, Bn |
AbsScalar(Vector64<Int64>) |
int64x1_t vabs_s64 (int64x1_t a) A64: ABS Dd, Dn |
Add(Vector128<Double>, Vector128<Double>) |
float64x2_t vaddq_f64 (float64x2_t a, float64x2_t b) A64: FADD Vd.2D, Vn.2D, Vm.2D |
AddAcross(Vector128<Byte>) |
uint8_t vaddvq_u8 (uint8x16_t a) A64: ADDV Bd, Vn.16B |
AddAcross(Vector128<Int16>) |
int16_t vaddvq_s16 (int16x8_t a) A64: ADDV Hd, Vn.8H |
AddAcross(Vector128<Int32>) |
int32_t vaddvq_s32 (int32x4_t a) A64: ADDV Sd, Vn.4S |
AddAcross(Vector128<SByte>) |
int8_t vaddvq_s8 (int8x16_t a) A64: ADDV Bd, Vn.16B |
AddAcross(Vector128<UInt16>) |
uint16_t vaddvq_u16 (uint16x8_t a) A64: ADDV Hd, Vn.8H |
AddAcross(Vector128<UInt32>) |
uint32_t vaddvq_u32 (uint32x4_t a) A64: ADDV Sd, Vn.4S |
AddAcross(Vector64<Byte>) |
uint8_t vaddv_u8 (uint8x8_t a) A64: ADDV Bd, Vn.8B |
AddAcross(Vector64<Int16>) |
int16_t vaddv_s16 (int16x4_t a) A64: ADDV Hd, Vn.4H |
AddAcross(Vector64<SByte>) |
int8_t vaddv_s8 (int8x8_t a) A64: ADDV Bd, Vn.8B |
AddAcross(Vector64<UInt16>) |
uint16_t vaddv_u16 (uint16x4_t a) A64: ADDV Hd, Vn.4H |
AddAcrossWidening(Vector128<Byte>) |
uint16_t vaddlvq_u8 (uint8x16_t a) A64: UADDLV Hd, Vn.16B |
AddAcrossWidening(Vector128<Int16>) |
int32_t vaddlvq_s16 (int16x8_t a) A64: SADDLV Sd, Vn.8H |
AddAcrossWidening(Vector128<Int32>) |
int64_t vaddlvq_s32 (int32x4_t a) A64: SADDLV Dd, Vn.4S |
AddAcrossWidening(Vector128<SByte>) |
int16_t vaddlvq_s8 (int8x16_t a) A64: SADDLV Hd, Vn.16B |
AddAcrossWidening(Vector128<UInt16>) |
uint32_t vaddlvq_u16 (uint16x8_t a) A64: UADDLV Sd, Vn.8H |
AddAcrossWidening(Vector128<UInt32>) |
uint64_t vaddlvq_u32 (uint32x4_t a) A64: UADDLV Dd, Vn.4S |
AddAcrossWidening(Vector64<Byte>) |
uint16_t vaddlv_u8 (uint8x8_t a) A64: UADDLV Hd, Vn.8B |
AddAcrossWidening(Vector64<Int16>) |
int32_t vaddlv_s16 (int16x4_t a) A64: SADDLV Sd, Vn.4H |
AddAcrossWidening(Vector64<SByte>) |
int16_t vaddlv_s8 (int8x8_t a) A64: SADDLV Hd, Vn.8B |
AddAcrossWidening(Vector64<UInt16>) |
uint32_t vaddlv_u16 (uint16x4_t a) A64: UADDLV Sd, Vn.4H |
AddPairwise(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vpaddq_u8 (uint8x16_t a, uint8x16_t b) A64: ADDP Vd.16B, Vn.16B, Vm.16B |
AddPairwise(Vector128<Double>, Vector128<Double>) |
float64x2_t vpaddq_f64 (float64x2_t a, float64x2_t b) A64: FADDP Vd.2D, Vn.2D, Vm.2D |
AddPairwise(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vpaddq_s16 (int16x8_t a, int16x8_t b) A64: ADDP Vd.8H, Vn.8H, Vm.8H |
AddPairwise(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vpaddq_s32 (int32x4_t a, int32x4_t b) A64: ADDP Vd.4S, Vn.4S, Vm.4S |
AddPairwise(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vpaddq_s64 (int64x2_t a, int64x2_t b) A64: ADDP Vd.2D, Vn.2D, Vm.2D |
AddPairwise(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vpaddq_s8 (int8x16_t a, int8x16_t b) A64: ADDP Vd.16B, Vn.16B, Vm.16B |
AddPairwise(Vector128<Single>, Vector128<Single>) |
float32x4_t vpaddq_f32 (float32x4_t a, float32x4_t b) A64: FADDP Vd.4S, Vn.4S, Vm.4S |
AddPairwise(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vpaddq_u16 (uint16x8_t a, uint16x8_t b) A64: ADDP Vd.8H, Vn.8H, Vm.8H |
AddPairwise(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vpaddq_u32 (uint32x4_t a, uint32x4_t b) A64: ADDP Vd.4S, Vn.4S, Vm.4S |
AddPairwise(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vpaddq_u64 (uint64x2_t a, uint64x2_t b) A64: ADDP Vd.2D, Vn.2D, Vm.2D |
AddPairwiseScalar(Vector128<Double>) |
float64_t vpaddd_f64 (float64x2_t a) A64: FADDP Dd, Vn.2D |
AddPairwiseScalar(Vector128<Int64>) |
int64_t vpaddd_s64 (int64x2_t a) A64: ADDP Dd, Vn.2D |
AddPairwiseScalar(Vector128<UInt64>) |
uint64_t vpaddd_u64 (uint64x2_t a) A64: ADDP Dd, Vn.2D |
AddPairwiseScalar(Vector64<Single>) |
float32_t vpadds_f32 (float32x2_t a) A64: FADDP Sd, Vn.2S |
AddSaturate(Vector128<Byte>, Vector128<SByte>) |
uint8x16_t vsqaddq_u8 (uint8x16_t a, int8x16_t b) A64: USQADD Vd.16B, Vn.16B |
AddSaturate(Vector128<Int16>, Vector128<UInt16>) |
int16x8_t vuqaddq_s16 (int16x8_t a, uint16x8_t b) A64: SUQADD Vd.8H, Vn.8H |
AddSaturate(Vector128<Int32>, Vector128<UInt32>) |
int32x4_t vuqaddq_s32 (int32x4_t a, uint32x4_t b) A64: SUQADD Vd.4S, Vn.4S |
AddSaturate(Vector128<Int64>, Vector128<UInt64>) |
int64x2_t vuqaddq_s64 (int64x2_t a, uint64x2_t b) A64: SUQADD Vd.2D, Vn.2D |
AddSaturate(Vector128<SByte>, Vector128<Byte>) |
int8x16_t vuqaddq_s8 (int8x16_t a, uint8x16_t b) A64: SUQADD Vd.16B, Vn.16B |
AddSaturate(Vector128<UInt16>, Vector128<Int16>) |
uint16x8_t vsqaddq_u16 (uint16x8_t a, int16x8_t b) A64: USQADD Vd.8H, Vn.8H |
AddSaturate(Vector128<UInt32>, Vector128<Int32>) |
uint32x4_t vsqaddq_u32 (uint32x4_t a, int32x4_t b) A64: USQADD Vd.4S, Vn.4S |
AddSaturate(Vector128<UInt64>, Vector128<Int64>) |
uint64x2_t vsqaddq_u64 (uint64x2_t a, int64x2_t b) A64: USQADD Vd.2D, Vn.2D |
AddSaturate(Vector64<Byte>, Vector64<SByte>) |
uint8x8_t vsqadd_u8 (uint8x8_t a, int8x8_t b) A64: USQADD Vd.8B, Vn.8B |
AddSaturate(Vector64<Int16>, Vector64<UInt16>) |
int16x4_t vuqadd_s16 (int16x4_t a, uint16x4_t b) A64: SUQADD Vd.4H, Vn.4H |
AddSaturate(Vector64<Int32>, Vector64<UInt32>) |
int32x2_t vuqadd_s32 (int32x2_t a, uint32x2_t b) A64: SUQADD Vd.2S, Vn.2S |
AddSaturate(Vector64<SByte>, Vector64<Byte>) |
int8x8_t vuqadd_s8 (int8x8_t a, uint8x8_t b) A64: SUQADD Vd.8B, Vn.8B |
AddSaturate(Vector64<UInt16>, Vector64<Int16>) |
uint16x4_t vsqadd_u16 (uint16x4_t a, int16x4_t b) A64: USQADD Vd.4H, Vn.4H |
AddSaturate(Vector64<UInt32>, Vector64<Int32>) |
uint32x2_t vsqadd_u32 (uint32x2_t a, int32x2_t b) A64: USQADD Vd.2S, Vn.2S |
AddSaturateScalar(Vector64<Byte>, Vector64<Byte>) |
uint8_t vqaddb_u8 (uint8_t a, uint8_t b) A64: UQADD Bd, Bn, Bm |
AddSaturateScalar(Vector64<Byte>, Vector64<SByte>) |
uint8_t vsqaddb_u8 (uint8_t a, int8_t b) A64: USQADD Bd, Bn |
AddSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqaddh_s16 (int16_t a, int16_t b) A64: SQADD Hd, Hn, Hm |
AddSaturateScalar(Vector64<Int16>, Vector64<UInt16>) |
int16_t vuqaddh_s16 (int16_t a, uint16_t b) A64: SUQADD Hd, Hn |
AddSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqadds_s32 (int32_t a, int32_t b) A64: SQADD Sd, Sn, Sm |
AddSaturateScalar(Vector64<Int32>, Vector64<UInt32>) |
int32_t vuqadds_s32 (int32_t a, uint32_t b) A64: SUQADD Sd, Sn |
AddSaturateScalar(Vector64<Int64>, Vector64<UInt64>) |
int64x1_t vuqadd_s64 (int64x1_t a, uint64x1_t b) A64: SUQADD Dd, Dn |
AddSaturateScalar(Vector64<SByte>, Vector64<Byte>) |
int8_t vuqaddb_s8 (int8_t a, uint8_t b) A64: SUQADD Bd, Bn |
AddSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
int8_t vqaddb_s8 (int8_t a, int8_t b) A64: SQADD Bd, Bn, Bm |
AddSaturateScalar(Vector64<UInt16>, Vector64<Int16>) |
uint16_t vsqaddh_u16 (uint16_t a, int16_t b) A64: USQADD Hd, Hn |
AddSaturateScalar(Vector64<UInt16>, Vector64<UInt16>) |
uint16_t vqaddh_u16 (uint16_t a, uint16_t b) A64: UQADD Hd, Hn, Hm |
AddSaturateScalar(Vector64<UInt32>, Vector64<Int32>) |
uint32_t vsqadds_u32 (uint32_t a, int32_t b) A64: USQADD Sd, Sn |
AddSaturateScalar(Vector64<UInt32>, Vector64<UInt32>) |
uint32_t vqadds_u32 (uint32_t a, uint32_t b) A64: UQADD Sd, Sn, Sm |
AddSaturateScalar(Vector64<UInt64>, Vector64<Int64>) |
uint64x1_t vsqadd_u64 (uint64x1_t a, int64x1_t b) A64: USQADD Dd, Dn |
Ceiling(Vector128<Double>) |
float64x2_t vrndpq_f64 (float64x2_t a) A64: FRINTP Vd.2D, Vn.2D |
CompareEqual(Vector128<Double>, Vector128<Double>) |
uint64x2_t vceqq_f64 (float64x2_t a, float64x2_t b) A64: FCMEQ Vd.2D, Vn.2D, Vm.2D |
CompareEqual(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vceqq_s64 (int64x2_t a, int64x2_t b) A64: CMEQ Vd.2D, Vn.2D, Vm.2D |
CompareEqual(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vceqq_u64 (uint64x2_t a, uint64x2_t b) A64: CMEQ Vd.2D, Vn.2D, Vm.2D |
CompareEqualScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vceq_f64 (float64x1_t a, float64x1_t b) A64: FCMEQ Dd, Dn, Dm |
CompareEqualScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vceq_s64 (int64x1_t a, int64x1_t b) A64: CMEQ Dd, Dn, Dm |
CompareEqualScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vceqs_f32 (float32_t a, float32_t b) A64: FCMEQ Sd, Sn, Sm |
CompareEqualScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vceq_u64 (uint64x1_t a, uint64x1_t b) A64: CMEQ Dd, Dn, Dm |
CompareGreaterThan(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcgtq_f64 (float64x2_t a, float64x2_t b) A64: FCMGT Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThan(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vcgtq_s64 (int64x2_t a, int64x2_t b) A64: CMGT Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThan(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vcgtq_u64 (uint64x2_t a, uint64x2_t b) A64: CMHI Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThanOrEqual(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcgeq_f64 (float64x2_t a, float64x2_t b) A64: FCMGE Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThanOrEqual(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vcgeq_s64 (int64x2_t a, int64x2_t b) A64: CMGE Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThanOrEqual(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vcgeq_u64 (uint64x2_t a, uint64x2_t b) A64: CMHS Vd.2D, Vn.2D, Vm.2D |
CompareGreaterThanOrEqualScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcge_f64 (float64x1_t a, float64x1_t b) A64: FCMGE Dd, Dn, Dm |
CompareGreaterThanOrEqualScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vcge_s64 (int64x1_t a, int64x1_t b) A64: CMGE Dd, Dn, Dm |
CompareGreaterThanOrEqualScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcges_f32 (float32_t a, float32_t b) A64: FCMGE Sd, Sn, Sm |
CompareGreaterThanOrEqualScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vcge_u64 (uint64x1_t a, uint64x1_t b) A64: CMHS Dd, Dn, Dm |
CompareGreaterThanScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcgt_f64 (float64x1_t a, float64x1_t b) A64: FCMGT Dd, Dn, Dm |
CompareGreaterThanScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vcgt_s64 (int64x1_t a, int64x1_t b) A64: CMGT Dd, Dn, Dm |
CompareGreaterThanScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcgts_f32 (float32_t a, float32_t b) A64: FCMGT Sd, Sn, Sm |
CompareGreaterThanScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vcgt_u64 (uint64x1_t a, uint64x1_t b) A64: CMHI Dd, Dn, Dm |
CompareLessThan(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcltq_f64 (float64x2_t a, float64x2_t b) A64: FCMGT Vd.2D, Vn.2D, Vm.2D |
CompareLessThan(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vcltq_s64 (int64x2_t a, int64x2_t b) A64: CMGT Vd.2D, Vn.2D, Vm.2D |
CompareLessThan(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vcltq_u64 (uint64x2_t a, uint64x2_t b) A64: CMHI Vd.2D, Vn.2D, Vm.2D |
CompareLessThanOrEqual(Vector128<Double>, Vector128<Double>) |
uint64x2_t vcleq_f64 (float64x2_t a, float64x2_t b) A64: FCMGE Vd.2D, Vn.2D, Vm.2D |
CompareLessThanOrEqual(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vcleq_s64 (int64x2_t a, int64x2_t b) A64: CMGE Vd.2D, Vn.2D, Vm.2D |
CompareLessThanOrEqual(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vcleq_u64 (uint64x2_t a, uint64x2_t b) A64: CMHS Vd.2D, Vn.2D, Vm.2D |
CompareLessThanOrEqualScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vcle_f64 (float64x1_t a, float64x1_t b) A64: FCMGE Dd, Dn, Dm |
CompareLessThanOrEqualScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vcle_s64 (int64x1_t a, int64x1_t b) A64: CMGE Dd, Dn, Dm |
CompareLessThanOrEqualScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vcles_f32 (float32_t a, float32_t b) A64: FCMGE Sd, Sn, Sm |
CompareLessThanOrEqualScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vcle_u64 (uint64x1_t a, uint64x1_t b) A64: CMHS Dd, Dn, Dm |
CompareLessThanScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vclt_f64 (float64x1_t a, float64x1_t b) A64: FCMGT Dd, Dn, Dm |
CompareLessThanScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vclt_s64 (int64x1_t a, int64x1_t b) A64: CMGT Dd, Dn, Dm |
CompareLessThanScalar(Vector64<Single>, Vector64<Single>) |
uint32_t vclts_f32 (float32_t a, float32_t b) A64: FCMGT Sd, Sn, Sm |
CompareLessThanScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vclt_u64 (uint64x1_t a, uint64x1_t b) A64: CMHI Dd, Dn, Dm |
CompareTest(Vector128<Double>, Vector128<Double>) |
uint64x2_t vtstq_f64 (float64x2_t a, float64x2_t b) A64: CMTST Vd.2D, Vn.2D, Vm.2D The above native signature does not exist. We provide this additional overload for consistency with the other scalar APIs. |
CompareTest(Vector128<Int64>, Vector128<Int64>) |
uint64x2_t vtstq_s64 (int64x2_t a, int64x2_t b) A64: CMTST Vd.2D, Vn.2D, Vm.2D |
CompareTest(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vtstq_u64 (uint64x2_t a, uint64x2_t b) A64: CMTST Vd.2D, Vn.2D, Vm.2D |
CompareTestScalar(Vector64<Double>, Vector64<Double>) |
uint64x1_t vtst_f64 (float64x1_t a, float64x1_t b) A64: CMTST Dd, Dn, Dm The above native signature does not exist. We provide this additional overload for consistency with the other scalar APIs. |
CompareTestScalar(Vector64<Int64>, Vector64<Int64>) |
uint64x1_t vtst_s64 (int64x1_t a, int64x1_t b) A64: CMTST Dd, Dn, Dm |
CompareTestScalar(Vector64<UInt64>, Vector64<UInt64>) |
uint64x1_t vtst_u64 (uint64x1_t a, uint64x1_t b) A64: CMTST Dd, Dn, Dm |
ConvertToDouble(Vector128<Int64>) |
float64x2_t vcvtq_f64_s64 (int64x2_t a) A64: SCVTF Vd.2D, Vn.2D |
ConvertToDouble(Vector128<UInt64>) |
float64x2_t vcvtq_f64_u64 (uint64x2_t a) A64: UCVTF Vd.2D, Vn.2D |
ConvertToDouble(Vector64<Single>) |
float64x2_t vcvt_f64_f32 (float32x2_t a) A64: FCVTL Vd.2D, Vn.2S |
ConvertToDoubleScalar(Vector64<Int64>) |
float64x1_t vcvt_f64_s64 (int64x1_t a) A64: SCVTF Dd, Dn |
ConvertToDoubleScalar(Vector64<UInt64>) |
float64x1_t vcvt_f64_u64 (uint64x1_t a) A64: UCVTF Dd, Dn |
ConvertToDoubleUpper(Vector128<Single>) |
float64x2_t vcvt_high_f64_f32 (float32x4_t a) A64: FCVTL2 Vd.2D, Vn.4S |
ConvertToInt64RoundAwayFromZero(Vector128<Double>) |
int64x2_t vcvtaq_s64_f64 (float64x2_t a) A64: FCVTAS Vd.2D, Vn.2D |
ConvertToInt64RoundAwayFromZeroScalar(Vector64<Double>) |
int64x1_t vcvta_s64_f64 (float64x1_t a) A64: FCVTAS Dd, Dn |
ConvertToInt64RoundToEven(Vector128<Double>) |
int64x2_t vcvtnq_s64_f64 (float64x2_t a) A64: FCVTNS Vd.2D, Vn.2D |
ConvertToInt64RoundToEvenScalar(Vector64<Double>) |
int64x1_t vcvtn_s64_f64 (float64x1_t a) A64: FCVTNS Dd, Dn |
ConvertToInt64RoundToNegativeInfinity(Vector128<Double>) |
int64x2_t vcvtmq_s64_f64 (float64x2_t a) A64: FCVTMS Vd.2D, Vn.2D |
ConvertToInt64RoundToNegativeInfinityScalar(Vector64<Double>) |
int64x1_t vcvtm_s64_f64 (float64x1_t a) A64: FCVTMS Dd, Dn |
ConvertToInt64RoundToPositiveInfinity(Vector128<Double>) |
int64x2_t vcvtpq_s64_f64 (float64x2_t a) A64: FCVTPS Vd.2D, Vn.2D |
ConvertToInt64RoundToPositiveInfinityScalar(Vector64<Double>) |
int64x1_t vcvtp_s64_f64 (float64x1_t a) A64: FCVTPS Dd, Dn |
ConvertToInt64RoundToZero(Vector128<Double>) |
int64x2_t vcvtq_s64_f64 (float64x2_t a) A64: FCVTZS Vd.2D, Vn.2D |
ConvertToInt64RoundToZeroScalar(Vector64<Double>) |
int64x1_t vcvt_s64_f64 (float64x1_t a) A64: FCVTZS Dd, Dn |
ConvertToSingleLower(Vector128<Double>) |
float32x2_t vcvt_f32_f64 (float64x2_t a) A64: FCVTN Vd.2S, Vn.2D |
ConvertToSingleRoundToOddLower(Vector128<Double>) |
float32x2_t vcvtx_f32_f64 (float64x2_t a) A64: FCVTXN Vd.2S, Vn.2D |
ConvertToSingleRoundToOddUpper(Vector64<Single>, Vector128<Double>) |
float32x4_t vcvtx_high_f32_f64 (float32x2_t r, float64x2_t a) A64: FCVTXN2 Vd.4S, Vn.2D |
ConvertToSingleUpper(Vector64<Single>, Vector128<Double>) |
float32x4_t vcvt_high_f32_f64 (float32x2_t r, float64x2_t a) A64: FCVTN2 Vd.4S, Vn.2D |
ConvertToUInt64RoundAwayFromZero(Vector128<Double>) |
uint64x2_t vcvtaq_u64_f64 (float64x2_t a) A64: FCVTAU Vd.2D, Vn.2D |
ConvertToUInt64RoundAwayFromZeroScalar(Vector64<Double>) |
uint64x1_t vcvta_u64_f64 (float64x1_t a) A64: FCVTAU Dd, Dn |
ConvertToUInt64RoundToEven(Vector128<Double>) |
uint64x2_t vcvtnq_u64_f64 (float64x2_t a) A64: FCVTNU Vd.2D, Vn.2D |
ConvertToUInt64RoundToEvenScalar(Vector64<Double>) |
uint64x1_t vcvtn_u64_f64 (float64x1_t a) A64: FCVTNU Dd, Dn |
ConvertToUInt64RoundToNegativeInfinity(Vector128<Double>) |
uint64x2_t vcvtmq_u64_f64 (float64x2_t a) A64: FCVTMU Vd.2D, Vn.2D |
ConvertToUInt64RoundToNegativeInfinityScalar(Vector64<Double>) |
uint64x1_t vcvtm_u64_f64 (float64x1_t a) A64: FCVTMU Dd, Dn |
ConvertToUInt64RoundToPositiveInfinity(Vector128<Double>) |
uint64x2_t vcvtpq_u64_f64 (float64x2_t a) A64: FCVTPU Vd.2D, Vn.2D |
ConvertToUInt64RoundToPositiveInfinityScalar(Vector64<Double>) |
uint64x1_t vcvtp_u64_f64 (float64x1_t a) A64: FCVTPU Dd, Dn |
ConvertToUInt64RoundToZero(Vector128<Double>) |
uint64x2_t vcvtq_u64_f64 (float64x2_t a) A64: FCVTZU Vd.2D, Vn.2D |
ConvertToUInt64RoundToZeroScalar(Vector64<Double>) |
uint64x1_t vcvt_u64_f64 (float64x1_t a) A64: FCVTZU Dd, Dn |
Divide(Vector128<Double>, Vector128<Double>) |
float64x2_t vdivq_f64 (float64x2_t a, float64x2_t b) A64: FDIV Vd.2D, Vn.2D, Vm.2D |
Divide(Vector128<Single>, Vector128<Single>) |
float32x4_t vdivq_f32 (float32x4_t a, float32x4_t b) A64: FDIV Vd.4S, Vn.4S, Vm.4S |
Divide(Vector64<Single>, Vector64<Single>) |
float32x2_t vdiv_f32 (float32x2_t a, float32x2_t b) A64: FDIV Vd.2S, Vn.2S, Vm.2S |
DuplicateSelectedScalarToVector128(Vector128<Double>, Byte) |
float64x2_t vdupq_laneq_f64 (float64x2_t vec, const int lane) A64: DUP Vd.2D, Vn.D[index] |
DuplicateSelectedScalarToVector128(Vector128<Int64>, Byte) |
int64x2_t vdupq_laneq_s64 (int64x2_t vec, const int lane) A64: DUP Vd.2D, Vn.D[index] |
DuplicateSelectedScalarToVector128(Vector128<UInt64>, Byte) |
uint64x2_t vdupq_laneq_u64 (uint64x2_t vec, const int lane) A64: DUP Vd.2D, Vn.D[index] |
DuplicateToVector128(Double) |
float64x2_t vdupq_n_f64 (float64_t value) A64: DUP Vd.2D, Vn.D[0] |
DuplicateToVector128(Int64) |
int64x2_t vdupq_n_s64 (int64_t value) A64: DUP Vd.2D, Rn |
DuplicateToVector128(UInt64) |
uint64x2_t vdupq_n_s64 (uint64_t value) A64: DUP Vd.2D, Rn |
Equals(Object) |
Determines whether the specified object is equal to the current object. (Inherited from Object) |
ExtractNarrowingSaturateScalar(Vector64<Int16>) |
int8_t vqmovnh_s16 (int16_t a) A64: SQXTN Bd, Hn |
ExtractNarrowingSaturateScalar(Vector64<Int32>) |
int16_t vqmovns_s32 (int32_t a) A64: SQXTN Hd, Sn |
ExtractNarrowingSaturateScalar(Vector64<Int64>) |
int32_t vqmovnd_s64 (int64_t a) A64: SQXTN Sd, Dn |
ExtractNarrowingSaturateScalar(Vector64<UInt16>) |
uint8_t vqmovnh_u16 (uint16_t a) A64: UQXTN Bd, Hn |
ExtractNarrowingSaturateScalar(Vector64<UInt32>) |
uint16_t vqmovns_u32 (uint32_t a) A64: UQXTN Hd, Sn |
ExtractNarrowingSaturateScalar(Vector64<UInt64>) |
uint32_t vqmovnd_u64 (uint64_t a) A64: UQXTN Sd, Dn |
ExtractNarrowingSaturateUnsignedScalar(Vector64<Int16>) |
uint8_t vqmovunh_s16 (int16_t a) A64: SQXTUN Bd, Hn |
ExtractNarrowingSaturateUnsignedScalar(Vector64<Int32>) |
uint16_t vqmovuns_s32 (int32_t a) A64: SQXTUN Hd, Sn |
ExtractNarrowingSaturateUnsignedScalar(Vector64<Int64>) |
uint32_t vqmovund_s64 (int64_t a) A64: SQXTUN Sd, Dn |
Floor(Vector128<Double>) |
float64x2_t vrndmq_f64 (float64x2_t a) A64: FRINTM Vd.2D, Vn.2D |
FusedMultiplyAdd(Vector128<Double>, Vector128<Double>, Vector128<Double>) |
float64x2_t vfmaq_f64 (float64x2_t a, float64x2_t b, float64x2_t c) A64: FMLA Vd.2D, Vn.2D, Vm.2D |
FusedMultiplyAddByScalar(Vector128<Double>, Vector128<Double>, Vector64<Double>) |
float64x2_t vfmaq_n_f64 (float64x2_t a, float64x2_t b, float64_t n) A64: FMLA Vd.2D, Vn.2D, Vm.D[0] |
FusedMultiplyAddByScalar(Vector128<Single>, Vector128<Single>, Vector64<Single>) |
float32x4_t vfmaq_n_f32 (float32x4_t a, float32x4_t b, float32_t n) A64: FMLA Vd.4S, Vn.4S, Vm.S[0] |
FusedMultiplyAddByScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>) |
float32x2_t vfma_n_f32 (float32x2_t a, float32x2_t b, float32_t n) A64: FMLA Vd.2S, Vn.2S, Vm.S[0] |
FusedMultiplyAddBySelectedScalar(Vector128<Double>, Vector128<Double>, Vector128<Double>, Byte) |
float64x2_t vfmaq_laneq_f64 (float64x2_t a, float64x2_t b, float64x2_t v, const int lane) A64: FMLA Vd.2D, Vn.2D, Vm.D[lane] |
FusedMultiplyAddBySelectedScalar(Vector128<Single>, Vector128<Single>, Vector128<Single>, Byte) |
float32x4_t vfmaq_laneq_f32 (float32x4_t a, float32x4_t b, float32x4_t v, const int lane) A64: FMLA Vd.4S, Vn.4S, Vm.S[lane] |
FusedMultiplyAddBySelectedScalar(Vector128<Single>, Vector128<Single>, Vector64<Single>, Byte) |
float32x4_t vfmaq_lane_f32 (float32x4_t a, float32x4_t b, float32x2_t v, const int lane) A64: FMLA Vd.4S, Vn.4S, Vm.S[lane] |
FusedMultiplyAddBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector128<Single>, Byte) |
float32x2_t vfma_laneq_f32 (float32x2_t a, float32x2_t b, float32x4_t v, const int lane) A64: FMLA Vd.2S, Vn.2S, Vm.S[lane] |
FusedMultiplyAddBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>, Byte) |
float32x2_t vfma_lane_f32 (float32x2_t a, float32x2_t b, float32x2_t v, const int lane) A64: FMLA Vd.2S, Vn.2S, Vm.S[lane] |
FusedMultiplyAddScalarBySelectedScalar(Vector64<Double>, Vector64<Double>, Vector128<Double>, Byte) |
float64_t vfmad_laneq_f64 (float64_t a, float64_t b, float64x2_t v, const int lane) A64: FMLA Dd, Dn, Vm.D[lane] |
FusedMultiplyAddScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector128<Single>, Byte) |
float32_t vfmas_laneq_f32 (float32_t a, float32_t b, float32x4_t v, const int lane) A64: FMLA Sd, Sn, Vm.S[lane] |
FusedMultiplyAddScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>, Byte) |
float32_t vfmas_lane_f32 (float32_t a, float32_t b, float32x2_t v, const int lane) A64: FMLA Sd, Sn, Vm.S[lane] |
FusedMultiplySubtract(Vector128<Double>, Vector128<Double>, Vector128<Double>) |
float64x2_t vfmsq_f64 (float64x2_t a, float64x2_t b, float64x2_t c) A64: FMLS Vd.2D, Vn.2D, Vm.2D |
FusedMultiplySubtractByScalar(Vector128<Double>, Vector128<Double>, Vector64<Double>) |
float64x2_t vfmsq_n_f64 (float64x2_t a, float64x2_t b, float64_t n) A64: FMLS Vd.2D, Vn.2D, Vm.D[0] |
FusedMultiplySubtractByScalar(Vector128<Single>, Vector128<Single>, Vector64<Single>) |
float32x4_t vfmsq_n_f32 (float32x4_t a, float32x4_t b, float32_t n) A64: FMLS Vd.4S, Vn.4S, Vm.S[0] |
FusedMultiplySubtractByScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>) |
float32x2_t vfms_n_f32 (float32x2_t a, float32x2_t b, float32_t n) A64: FMLS Vd.2S, Vn.2S, Vm.S[0] |
FusedMultiplySubtractBySelectedScalar(Vector128<Double>, Vector128<Double>, Vector128<Double>, Byte) |
float64x2_t vfmsq_laneq_f64 (float64x2_t a, float64x2_t b, float64x2_t v, const int lane) A64: FMLS Vd.2D, Vn.2D, Vm.D[lane] |
FusedMultiplySubtractBySelectedScalar(Vector128<Single>, Vector128<Single>, Vector128<Single>, Byte) |
float32x4_t vfmsq_laneq_f32 (float32x4_t a, float32x4_t b, float32x4_t v, const int lane) A64: FMLS Vd.4S, Vn.4S, Vm.S[lane] |
FusedMultiplySubtractBySelectedScalar(Vector128<Single>, Vector128<Single>, Vector64<Single>, Byte) |
float32x4_t vfmsq_lane_f32 (float32x4_t a, float32x4_t b, float32x2_t v, const int lane) A64: FMLS Vd.4S, Vn.4S, Vm.S[lane] |
FusedMultiplySubtractBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector128<Single>, Byte) |
float32x2_t vfms_laneq_f32 (float32x2_t a, float32x2_t b, float32x4_t v, const int lane) A64: FMLS Vd.2S, Vn.2S, Vm.S[lane] |
FusedMultiplySubtractBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>, Byte) |
float32x2_t vfms_lane_f32 (float32x2_t a, float32x2_t b, float32x2_t v, const int lane) A64: FMLS Vd.2S, Vn.2S, Vm.S[lane] |
FusedMultiplySubtractScalarBySelectedScalar(Vector64<Double>, Vector64<Double>, Vector128<Double>, Byte) |
float64_t vfmsd_laneq_f64 (float64_t a, float64_t b, float64x2_t v, const int lane) A64: FMLS Dd, Dn, Vm.D[lane] |
FusedMultiplySubtractScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector128<Single>, Byte) |
float32_t vfmss_laneq_f32 (float32_t a, float32_t b, float32x4_t v, const int lane) A64: FMLS Sd, Sn, Vm.S[lane] |
FusedMultiplySubtractScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>, Byte) |
float32_t vfmss_lane_f32 (float32_t a, float32_t b, float32x2_t v, const int lane) A64: FMLS Sd, Sn, Vm.S[lane] |
GetHashCode() |
Serves as the default hash function. (Inherited from Object) |
GetType() |
Gets the Type of the current instance. (Inherited from Object) |
InsertSelectedScalar(Vector128<Byte>, Byte, Vector128<Byte>, Byte) |
uint8x16_t vcopyq_laneq_u8 (uint8x16_t a, const int lane1, uint8x16_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector128<Byte>, Byte, Vector64<Byte>, Byte) |
uint8x16_t vcopyq_lane_u8 (uint8x16_t a, const int lane1, uint8x8_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector128<Double>, Byte, Vector128<Double>, Byte) |
float64x2_t vcopyq_laneq_f64 (float64x2_t a, const int lane1, float64x2_t b, const int lane2) A64: INS Vd.D[lane1], Vn.D[lane2] |
InsertSelectedScalar(Vector128<Int16>, Byte, Vector128<Int16>, Byte) |
int16x8_t vcopyq_laneq_s16 (int16x8_t a, const int lane1, int16x8_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector128<Int16>, Byte, Vector64<Int16>, Byte) |
int16x8_t vcopyq_lane_s16 (int16x8_t a, const int lane1, int16x4_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector128<Int32>, Byte, Vector128<Int32>, Byte) |
int32x4_t vcopyq_laneq_s32 (int32x4_t a, const int lane1, int32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<Int32>, Byte, Vector64<Int32>, Byte) |
int32x4_t vcopyq_lane_s32 (int32x4_t a, const int lane1, int32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<Int64>, Byte, Vector128<Int64>, Byte) |
int64x2_t vcopyq_laneq_s64 (int64x2_t a, const int lane1, int64x2_t b, const int lane2) A64: INS Vd.D[lane1], Vn.D[lane2] |
InsertSelectedScalar(Vector128<SByte>, Byte, Vector128<SByte>, Byte) |
int8x16_t vcopyq_laneq_s8 (int8x16_t a, const int lane1, int8x16_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector128<SByte>, Byte, Vector64<SByte>, Byte) |
int8x16_t vcopyq_lane_s8 (int8x16_t a, const int lane1, int8x8_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector128<Single>, Byte, Vector128<Single>, Byte) |
float32x4_t vcopyq_laneq_f32 (float32x4_t a, const int lane1, float32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<Single>, Byte, Vector64<Single>, Byte) |
float32x4_t vcopyq_lane_f32 (float32x4_t a, const int lane1, float32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<UInt16>, Byte, Vector128<UInt16>, Byte) |
uint16x8_t vcopyq_laneq_u16 (uint16x8_t a, const int lane1, uint16x8_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector128<UInt16>, Byte, Vector64<UInt16>, Byte) |
uint16x8_t vcopyq_lane_u16 (uint16x8_t a, const int lane1, uint16x4_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector128<UInt32>, Byte, Vector128<UInt32>, Byte) |
uint32x4_t vcopyq_laneq_u32 (uint32x4_t a, const int lane1, uint32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<UInt32>, Byte, Vector64<UInt32>, Byte) |
uint32x4_t vcopyq_lane_u32 (uint32x4_t a, const int lane1, uint32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector128<UInt64>, Byte, Vector128<UInt64>, Byte) |
uint64x2_t vcopyq_laneq_u64 (uint64x2_t a, const int lane1, uint64x2_t b, const int lane2) A64: INS Vd.D[lane1], Vn.D[lane2] |
InsertSelectedScalar(Vector64<Byte>, Byte, Vector128<Byte>, Byte) |
uint8x8_t vcopy_laneq_u8 (uint8x8_t a, const int lane1, uint8x16_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector64<Byte>, Byte, Vector64<Byte>, Byte) |
uint8x8_t vcopy_lane_u8 (uint8x8_t a, const int lane1, uint8x8_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector64<Int16>, Byte, Vector128<Int16>, Byte) |
int16x4_t vcopy_laneq_s16 (int16x4_t a, const int lane1, int16x8_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector64<Int16>, Byte, Vector64<Int16>, Byte) |
int16x4_t vcopy_lane_s16 (int16x4_t a, const int lane1, int16x4_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector64<Int32>, Byte, Vector128<Int32>, Byte) |
int32x2_t vcopy_laneq_s32 (int32x2_t a, const int lane1, int32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector64<Int32>, Byte, Vector64<Int32>, Byte) |
int32x2_t vcopy_lane_s32 (int32x2_t a, const int lane1, int32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector64<SByte>, Byte, Vector128<SByte>, Byte) |
int8x8_t vcopy_laneq_s8 (int8x8_t a, const int lane1, int8x16_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector64<SByte>, Byte, Vector64<SByte>, Byte) |
int8x8_t vcopy_lane_s8 (int8x8_t a, const int lane1, int8x8_t b, const int lane2) A64: INS Vd.B[lane1], Vn.B[lane2] |
InsertSelectedScalar(Vector64<Single>, Byte, Vector128<Single>, Byte) |
float32x2_t vcopy_laneq_f32 (float32x2_t a, const int lane1, float32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector64<Single>, Byte, Vector64<Single>, Byte) |
float32x2_t vcopy_lane_f32 (float32x2_t a, const int lane1, float32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector64<UInt16>, Byte, Vector128<UInt16>, Byte) |
uint16x4_t vcopy_laneq_u16 (uint16x4_t a, const int lane1, uint16x8_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector64<UInt16>, Byte, Vector64<UInt16>, Byte) |
uint16x4_t vcopy_lane_u16 (uint16x4_t a, const int lane1, uint16x4_t b, const int lane2) A64: INS Vd.H[lane1], Vn.H[lane2] |
InsertSelectedScalar(Vector64<UInt32>, Byte, Vector128<UInt32>, Byte) |
uint32x2_t vcopy_laneq_u32 (uint32x2_t a, const int lane1, uint32x4_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
InsertSelectedScalar(Vector64<UInt32>, Byte, Vector64<UInt32>, Byte) |
uint32x2_t vcopy_lane_u32 (uint32x2_t a, const int lane1, uint32x2_t b, const int lane2) A64: INS Vd.S[lane1], Vn.S[lane2] |
Load2xVector128(Byte*) |
A64: LD1 { Vn.16B, Vn+1.16B }, [Xn] |
Load2xVector128(Double*) |
A64: LD1 { Vn.2D, Vn+1.2D }, [Xn] |
Load2xVector128(Int16*) |
A64: LD1 { Vn.8H, Vn+1.8H }, [Xn] |
Load2xVector128(Int32*) |
A64: LD1 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128(Int64*) |
A64: LD1 { Vn.2D, Vn+1.2D }, [Xn] |
Load2xVector128(SByte*) |
A64: LD1 { Vn.16B, Vn+1.16B }, [Xn] |
Load2xVector128(Single*) |
A64: LD1 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128(UInt16*) |
A64: LD1 { Vn.8H, Vn+1.8H }, [Xn] |
Load2xVector128(UInt32*) |
A64: LD1 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128(UInt64*) |
A64: LD1 { Vn.2D, Vn+1.2D }, [Xn] |
Load2xVector128AndUnzip(Byte*) |
A64: LD2 { Vn.16B, Vn+1.16B }, [Xn] |
Load2xVector128AndUnzip(Double*) |
A64: LD2 { Vn.2D, Vn+1.2D }, [Xn] |
Load2xVector128AndUnzip(Int16*) |
A64: LD2 { Vn.8H, Vn+1.8H }, [Xn] |
Load2xVector128AndUnzip(Int32*) |
A64: LD2 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128AndUnzip(Int64*) |
A64: LD2 { Vn.2D, Vn+1.2D }, [Xn] |
Load2xVector128AndUnzip(SByte*) |
A64: LD2 { Vn.16B, Vn+1.16B }, [Xn] |
Load2xVector128AndUnzip(Single*) |
A64: LD2 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128AndUnzip(UInt16*) |
A64: LD2 { Vn.8H, Vn+1.8H }, [Xn] |
Load2xVector128AndUnzip(UInt32*) |
A64: LD2 { Vn.4S, Vn+1.4S }, [Xn] |
Load2xVector128AndUnzip(UInt64*) |
A64: LD2 { Vn.2D, Vn+1.2D }, [Xn] |
Load3xVector128(Byte*) |
A64: LD1 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Load3xVector128(Double*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load3xVector128(Int16*) |
A64: LD1 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Load3xVector128(Int32*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128(Int64*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load3xVector128(SByte*) |
A64: LD1 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Load3xVector128(Single*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128(UInt16*) |
A64: LD1 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Load3xVector128(UInt32*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128(UInt64*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load3xVector128AndUnzip(Byte*) |
A64: LD3 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Load3xVector128AndUnzip(Double*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load3xVector128AndUnzip(Int16*) |
A64: LD3 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Load3xVector128AndUnzip(Int32*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128AndUnzip(Int64*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load3xVector128AndUnzip(SByte*) |
A64: LD3 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Load3xVector128AndUnzip(Single*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128AndUnzip(UInt16*) |
A64: LD3 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Load3xVector128AndUnzip(UInt32*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Load3xVector128AndUnzip(UInt64*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Load4xVector128(Byte*) |
A64: LD1 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Load4xVector128(Double*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Load4xVector128(Int16*) |
A64: LD1 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Load4xVector128(Int32*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128(Int64*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Load4xVector128(SByte*) |
A64: LD1 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Load4xVector128(Single*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128(UInt16*) |
A64: LD1 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Load4xVector128(UInt32*) |
A64: LD1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128(UInt64*) |
A64: LD1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D}, [Xn] |
Load4xVector128AndUnzip(Byte*) |
A64: LD4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Load4xVector128AndUnzip(Double*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Load4xVector128AndUnzip(Int16*) |
A64: LD4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Load4xVector128AndUnzip(Int32*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128AndUnzip(Int64*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Load4xVector128AndUnzip(SByte*) |
A64: LD4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Load4xVector128AndUnzip(Single*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128AndUnzip(UInt16*) |
A64: LD4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Load4xVector128AndUnzip(UInt32*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Load4xVector128AndUnzip(UInt64*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Byte, Byte*) |
A64: LD4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Byte, Byte*) |
A64: LD3 { Vn.16B, Vn+1.16B, Vn+2.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Byte>,Vector128<Byte>>, Byte, Byte*) |
A64: LD2 { Vn.16B, Vn+1.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>,Vector128<Double>>, Byte, Double*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>>, Byte, Double*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Double>,Vector128<Double>>, Byte, Double*) |
A64: LD2 { Vn.2D, Vn+1.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>, Byte, Int16*) |
A64: LD4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>, Byte, Int16*) |
A64: LD3 { Vn.8H, Vn+1.8H, Vn+2.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int16>,Vector128<Int16>>, Byte, Int16*) |
A64: LD2 { Vn.8H, Vn+1.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>, Byte, Int32*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>, Byte, Int32*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int32>,Vector128<Int32>>, Byte, Int32*) |
A64: LD2 { Vn.4S, Vn+1.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>, Byte, Int64*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>, Byte, Int64*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Int64>,Vector128<Int64>>, Byte, Int64*) |
A64: LD2 { Vn.2D, Vn+1.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Byte, SByte*) |
A64: LD4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Byte, SByte*) |
A64: LD3 { Vn.16B, Vn+1.16B, Vn+2.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<SByte>,Vector128<SByte>>, Byte, SByte*) |
A64: LD2 { Vn.16B, Vn+1.16B }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>,Vector128<Single>>, Byte, Single*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>>, Byte, Single*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<Single>,Vector128<Single>>, Byte, Single*) |
A64: LD2 { Vn.4S, Vn+1.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>, Byte, UInt16*) |
A64: LD4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>, Byte, UInt16*) |
A64: LD3 { Vn.8H, Vn+1.8H, Vn+2.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt16>,Vector128<UInt16>>, Byte, UInt16*) |
A64: LD2 { Vn.8H, Vn+1.8H }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>, Byte, UInt32*) |
A64: LD4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>, Byte, UInt32*) |
A64: LD3 { Vn.4S, Vn+1.4S, Vn+2.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt32>,Vector128<UInt32>>, Byte, UInt32*) |
A64: LD2 { Vn.4S, Vn+1.4S }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>, Byte, UInt64*) |
A64: LD4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>, Byte, UInt64*) |
A64: LD3 { Vn.2D, Vn+1.2D, Vn+2.2D }[Vm], [Xn] |
LoadAndInsertScalar(ValueTuple<Vector128<UInt64>,Vector128<UInt64>>, Byte, UInt64*) |
A64: LD2 { Vn.2D, Vn+1.2D }[Vm], [Xn] |
LoadAndReplicateToVector128(Double*) |
float64x2_t vld1q_dup_f64 (float64_t const * ptr) A64: LD1R { Vt.2D }, [Xn] |
LoadAndReplicateToVector128(Int64*) |
int64x2_t vld1q_dup_s64 (int64_t const * ptr) A64: LD1R { Vt.2D }, [Xn] |
LoadAndReplicateToVector128(UInt64*) |
uint64x2_t vld1q_dup_u64 (uint64_t const * ptr) A64: LD1R { Vt.2D }, [Xn] |
LoadAndReplicateToVector128x2(Byte*) |
A64: LD2R { Vn.16B, Vn+1.16B }, [Xn] |
LoadAndReplicateToVector128x2(Double*) |
A64: LD2R { Vn.2D, Vn+1.2D }, [Xn] |
LoadAndReplicateToVector128x2(Int16*) |
A64: LD2R { Vn.8H, Vn+1.8H }, [Xn] |
LoadAndReplicateToVector128x2(Int32*) |
A64: LD2R { Vn.4S, Vn+1.4S }, [Xn] |
LoadAndReplicateToVector128x2(Int64*) |
A64: LD2R { Vn.2D, Vn+1.2D }, [Xn] |
LoadAndReplicateToVector128x2(SByte*) |
A64: LD2R { Vn.16B, Vn+1.16B }, [Xn] |
LoadAndReplicateToVector128x2(Single*) |
A64: LD2R { Vn.4S, Vn+1.4S }, [Xn] |
LoadAndReplicateToVector128x2(UInt16*) |
A64: LD2R { Vn.8H, Vn+1.8H }, [Xn] |
LoadAndReplicateToVector128x2(UInt32*) |
A64: LD2R { Vn.4S, Vn+1.4S }, [Xn] |
LoadAndReplicateToVector128x2(UInt64*) |
A64: LD2R { Vn.2D, Vn+1.2D }, [Xn] |
LoadAndReplicateToVector128x3(Byte*) |
A64: LD3R { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
LoadAndReplicateToVector128x3(Double*) |
A64: LD3R { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
LoadAndReplicateToVector128x3(Int16*) |
A64: LD3R { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
LoadAndReplicateToVector128x3(Int32*) |
A64: LD3R { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
LoadAndReplicateToVector128x3(Int64*) |
A64: LD3R { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
LoadAndReplicateToVector128x3(SByte*) |
A64: LD3R { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
LoadAndReplicateToVector128x3(Single*) |
A64: LD3R { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
LoadAndReplicateToVector128x3(UInt16*) |
A64: LD3R { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
LoadAndReplicateToVector128x3(UInt32*) |
A64: LD3R { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
LoadAndReplicateToVector128x3(UInt64*) |
A64: LD3R { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
LoadAndReplicateToVector128x4(Byte*) |
A64: LD4R { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
LoadAndReplicateToVector128x4(Double*) |
A64: LD4R { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
LoadAndReplicateToVector128x4(Int16*) |
A64: LD4R { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
LoadAndReplicateToVector128x4(Int32*) |
A64: LD4R { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
LoadAndReplicateToVector128x4(Int64*) |
A64: LD4R { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
LoadAndReplicateToVector128x4(SByte*) |
A64: LD4R { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
LoadAndReplicateToVector128x4(Single*) |
A64: LD4R { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
LoadAndReplicateToVector128x4(UInt16*) |
A64: LD4R { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
LoadAndReplicateToVector128x4(UInt32*) |
A64: LD4R { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
LoadAndReplicateToVector128x4(UInt64*) |
A64: LD4R { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
LoadPairScalarVector64(Int32*) |
A64: LDP St1, St2, [Xn] |
LoadPairScalarVector64(Single*) |
A64: LDP St1, St2, [Xn] |
LoadPairScalarVector64(UInt32*) |
A64: LDP St1, St2, [Xn] |
LoadPairScalarVector64NonTemporal(Int32*) |
A64: LDNP St1, St2, [Xn] |
LoadPairScalarVector64NonTemporal(Single*) |
A64: LDNP St1, St2, [Xn] |
LoadPairScalarVector64NonTemporal(UInt32*) |
A64: LDNP St1, St2, [Xn] |
LoadPairVector128(Byte*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(Double*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(Int16*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(Int32*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(Int64*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(SByte*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(Single*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(UInt16*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(UInt32*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128(UInt64*) |
A64: LDP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Byte*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Double*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Int16*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Int32*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Int64*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(SByte*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(Single*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(UInt16*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(UInt32*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector128NonTemporal(UInt64*) |
A64: LDNP Qt1, Qt2, [Xn] |
LoadPairVector64(Byte*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(Double*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(Int16*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(Int32*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(Int64*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(SByte*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(Single*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(UInt16*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(UInt32*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64(UInt64*) |
A64: LDP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Byte*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Double*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Int16*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Int32*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Int64*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(SByte*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(Single*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(UInt16*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(UInt32*) |
A64: LDNP Dt1, Dt2, [Xn] |
LoadPairVector64NonTemporal(UInt64*) |
A64: LDNP Dt1, Dt2, [Xn] |
Max(Vector128<Double>, Vector128<Double>) |
float64x2_t vmaxq_f64 (float64x2_t a, float64x2_t b) A64: FMAX Vd.2D, Vn.2D, Vm.2D |
MaxAcross(Vector128<Byte>) |
uint8_t vmaxvq_u8 (uint8x16_t a) A64: UMAXV Bd, Vn.16B |
MaxAcross(Vector128<Int16>) |
int16_t vmaxvq_s16 (int16x8_t a) A64: SMAXV Hd, Vn.8H |
MaxAcross(Vector128<Int32>) |
int32_t vmaxvq_s32 (int32x4_t a) A64: SMAXV Sd, Vn.4S |
MaxAcross(Vector128<SByte>) |
int8_t vmaxvq_s8 (int8x16_t a) A64: SMAXV Bd, Vn.16B |
MaxAcross(Vector128<Single>) |
float32_t vmaxvq_f32 (float32x4_t a) A64: FMAXV Sd, Vn.4S |
MaxAcross(Vector128<UInt16>) |
uint16_t vmaxvq_u16 (uint16x8_t a) A64: UMAXV Hd, Vn.8H |
MaxAcross(Vector128<UInt32>) |
uint32_t vmaxvq_u32 (uint32x4_t a) A64: UMAXV Sd, Vn.4S |
MaxAcross(Vector64<Byte>) |
uint8_t vmaxv_u8 (uint8x8_t a) A64: UMAXV Bd, Vn.8B |
MaxAcross(Vector64<Int16>) |
int16_t vmaxv_s16 (int16x4_t a) A64: SMAXV Hd, Vn.4H |
MaxAcross(Vector64<SByte>) |
int8_t vmaxv_s8 (int8x8_t a) A64: SMAXV Bd, Vn.8B |
MaxAcross(Vector64<UInt16>) |
uint16_t vmaxv_u16 (uint16x4_t a) A64: UMAXV Hd, Vn.4H |
MaxNumber(Vector128<Double>, Vector128<Double>) |
float64x2_t vmaxnmq_f64 (float64x2_t a, float64x2_t b) A64: FMAXNM Vd.2D, Vn.2D, Vm.2D |
MaxNumberAcross(Vector128<Single>) |
float32_t vmaxnmvq_f32 (float32x4_t a) A64: FMAXNMV Sd, Vn.4S |
MaxNumberPairwise(Vector128<Double>, Vector128<Double>) |
float64x2_t vpmaxnmq_f64 (float64x2_t a, float64x2_t b) A64: FMAXNMP Vd.2D, Vn.2D, Vm.2D |
MaxNumberPairwise(Vector128<Single>, Vector128<Single>) |
float32x4_t vpmaxnmq_f32 (float32x4_t a, float32x4_t b) A64: FMAXNMP Vd.4S, Vn.4S, Vm.4S |
MaxNumberPairwise(Vector64<Single>, Vector64<Single>) |
float32x2_t vpmaxnm_f32 (float32x2_t a, float32x2_t b) A64: FMAXNMP Vd.2S, Vn.2S, Vm.2S |
MaxNumberPairwiseScalar(Vector128<Double>) |
float64_t vpmaxnmqd_f64 (float64x2_t a) A64: FMAXNMP Dd, Vn.2D |
MaxNumberPairwiseScalar(Vector64<Single>) |
float32_t vpmaxnms_f32 (float32x2_t a) A64: FMAXNMP Sd, Vn.2S |
MaxPairwise(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vpmaxq_u8 (uint8x16_t a, uint8x16_t b) A64: UMAXP Vd.16B, Vn.16B, Vm.16B |
MaxPairwise(Vector128<Double>, Vector128<Double>) |
float64x2_t vpmaxq_f64 (float64x2_t a, float64x2_t b) A64: FMAXP Vd.2D, Vn.2D, Vm.2D |
MaxPairwise(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vpmaxq_s16 (int16x8_t a, int16x8_t b) A64: SMAXP Vd.8H, Vn.8H, Vm.8H |
MaxPairwise(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vpmaxq_s32 (int32x4_t a, int32x4_t b) A64: SMAXP Vd.4S, Vn.4S, Vm.4S |
MaxPairwise(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vpmaxq_s8 (int8x16_t a, int8x16_t b) A64: SMAXP Vd.16B, Vn.16B, Vm.16B |
MaxPairwise(Vector128<Single>, Vector128<Single>) |
float32x4_t vpmaxq_f32 (float32x4_t a, float32x4_t b) A64: FMAXP Vd.4S, Vn.4S, Vm.4S |
MaxPairwise(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vpmaxq_u16 (uint16x8_t a, uint16x8_t b) A64: UMAXP Vd.8H, Vn.8H, Vm.8H |
MaxPairwise(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vpmaxq_u32 (uint32x4_t a, uint32x4_t b) A64: UMAXP Vd.4S, Vn.4S, Vm.4S |
MaxPairwiseScalar(Vector128<Double>) |
float64_t vpmaxqd_f64 (float64x2_t a) A64: FMAXP Dd, Vn.2D |
MaxPairwiseScalar(Vector64<Single>) |
float32_t vpmaxs_f32 (float32x2_t a) A64: FMAXP Sd, Vn.2S |
MaxScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vmax_f64 (float64x1_t a, float64x1_t b) A64: FMAX Dd, Dn, Dm |
MaxScalar(Vector64<Single>, Vector64<Single>) |
float32_t vmaxs_f32 (float32_t a, float32_t b) A64: FMAX Sd, Sn, Sm The above native signature does not exist. We provide this additional overload for consistency with the other scalar APIs. |
MemberwiseClone() |
Creates a shallow copy of the current Object. (Inherited from Object) |
Min(Vector128<Double>, Vector128<Double>) |
float64x2_t vminq_f64 (float64x2_t a, float64x2_t b) A64: FMIN Vd.2D, Vn.2D, Vm.2D |
MinAcross(Vector128<Byte>) |
uint8_t vminvq_u8 (uint8x16_t a) A64: UMINV Bd, Vn.16B |
MinAcross(Vector128<Int16>) |
int16_t vminvq_s16 (int16x8_t a) A64: SMINV Hd, Vn.8H |
MinAcross(Vector128<Int32>) |
int32_t vaddvq_s32 (int32x4_t a) A64: SMINV Sd, Vn.4S |
MinAcross(Vector128<SByte>) |
int8_t vminvq_s8 (int8x16_t a) A64: SMINV Bd, Vn.16B |
MinAcross(Vector128<Single>) |
float32_t vminvq_f32 (float32x4_t a) A64: FMINV Sd, Vn.4S |
MinAcross(Vector128<UInt16>) |
uint16_t vminvq_u16 (uint16x8_t a) A64: UMINV Hd, Vn.8H |
MinAcross(Vector128<UInt32>) |
uint32_t vminvq_u32 (uint32x4_t a) A64: UMINV Sd, Vn.4S |
MinAcross(Vector64<Byte>) |
uint8_t vminv_u8 (uint8x8_t a) A64: UMINV Bd, Vn.8B |
MinAcross(Vector64<Int16>) |
int16_t vminv_s16 (int16x4_t a) A64: SMINV Hd, Vn.4H |
MinAcross(Vector64<SByte>) |
int8_t vminv_s8 (int8x8_t a) A64: SMINV Bd, Vn.8B |
MinAcross(Vector64<UInt16>) |
uint16_t vminv_u16 (uint16x4_t a) A64: UMINV Hd, Vn.4H |
MinNumber(Vector128<Double>, Vector128<Double>) |
float64x2_t vminnmq_f64 (float64x2_t a, float64x2_t b) A64: FMINNM Vd.2D, Vn.2D, Vm.2D |
MinNumberAcross(Vector128<Single>) |
float32_t vminnmvq_f32 (float32x4_t a) A64: FMINNMV Sd, Vn.4S |
MinNumberPairwise(Vector128<Double>, Vector128<Double>) |
float64x2_t vpminnmq_f64 (float64x2_t a, float64x2_t b) A64: FMINNMP Vd.2D, Vn.2D, Vm.2D |
MinNumberPairwise(Vector128<Single>, Vector128<Single>) |
float32x4_t vpminnmq_f32 (float32x4_t a, float32x4_t b) A64: FMINNMP Vd.4S, Vn.4S, Vm.4S |
MinNumberPairwise(Vector64<Single>, Vector64<Single>) |
float32x2_t vpminnm_f32 (float32x2_t a, float32x2_t b) A64: FMINNMP Vd.2S, Vn.2S, Vm.2S |
MinNumberPairwiseScalar(Vector128<Double>) |
float64_t vpminnmqd_f64 (float64x2_t a) A64: FMINNMP Dd, Vn.2D |
MinNumberPairwiseScalar(Vector64<Single>) |
float32_t vpminnms_f32 (float32x2_t a) A64: FMINNMP Sd, Vn.2S |
MinPairwise(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vpminq_u8 (uint8x16_t a, uint8x16_t b) A64: UMINP Vd.16B, Vn.16B, Vm.16B |
MinPairwise(Vector128<Double>, Vector128<Double>) |
float64x2_t vpminq_f64 (float64x2_t a, float64x2_t b) A64: FMINP Vd.2D, Vn.2D, Vm.2D |
MinPairwise(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vpminq_s16 (int16x8_t a, int16x8_t b) A64: SMINP Vd.8H, Vn.8H, Vm.8H |
MinPairwise(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vpminq_s32 (int32x4_t a, int32x4_t b) A64: SMINP Vd.4S, Vn.4S, Vm.4S |
MinPairwise(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vpminq_s8 (int8x16_t a, int8x16_t b) A64: SMINP Vd.16B, Vn.16B, Vm.16B |
MinPairwise(Vector128<Single>, Vector128<Single>) |
float32x4_t vpminq_f32 (float32x4_t a, float32x4_t b) A64: FMINP Vd.4S, Vn.4S, Vm.4S |
MinPairwise(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vpminq_u16 (uint16x8_t a, uint16x8_t b) A64: UMINP Vd.8H, Vn.8H, Vm.8H |
MinPairwise(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vpminq_u32 (uint32x4_t a, uint32x4_t b) A64: UMINP Vd.4S, Vn.4S, Vm.4S |
MinPairwiseScalar(Vector128<Double>) |
float64_t vpminqd_f64 (float64x2_t a) A64: FMINP Dd, Vn.2D |
MinPairwiseScalar(Vector64<Single>) |
float32_t vpmins_f32 (float32x2_t a) A64: FMINP Sd, Vn.2S |
MinScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vmin_f64 (float64x1_t a, float64x1_t b) A64: FMIN Dd, Dn, Dm |
MinScalar(Vector64<Single>, Vector64<Single>) |
float32_t vmins_f32 (float32_t a, float32_t b) A64: FMIN Sd, Sn, Sm The above native signature does not exist. We provide this additional overload for consistency with the other scalar APIs. |
Multiply(Vector128<Double>, Vector128<Double>) |
float64x2_t vmulq_f64 (float64x2_t a, float64x2_t b) A64: FMUL Vd.2D, Vn.2D, Vm.2D |
MultiplyByScalar(Vector128<Double>, Vector64<Double>) |
float64x2_t vmulq_n_f64 (float64x2_t a, float64_t b) A64: FMUL Vd.2D, Vn.2D, Vm.D[0] |
MultiplyBySelectedScalar(Vector128<Double>, Vector128<Double>, Byte) |
float64x2_t vmulq_laneq_f64 (float64x2_t a, float64x2_t v, const int lane) A64: FMUL Vd.2D, Vn.2D, Vm.D[lane] |
MultiplyDoublingSaturateHighScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqdmulhh_s16 (int16_t a, int16_t b) A64: SQDMULH Hd, Hn, Hm |
MultiplyDoublingSaturateHighScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqdmulhs_s32 (int32_t a, int32_t b) A64: SQDMULH Sd, Sn, Sm |
MultiplyDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int16>, Vector128<Int16>, Byte) |
int16_t vqdmulhh_laneq_s16 (int16_t a, int16x8_t v, const int lane) A64: SQDMULH Hd, Hn, Vm.H[lane] |
MultiplyDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>, Byte) |
int16_t vqdmulhh_lane_s16 (int16_t a, int16x4_t v, const int lane) A64: SQDMULH Hd, Hn, Vm.H[lane] |
MultiplyDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int32>, Vector128<Int32>, Byte) |
int32_t vqdmulhs_laneq_s32 (int32_t a, int32x4_t v, const int lane) A64: SQDMULH Sd, Sn, Vm.S[lane] |
MultiplyDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>, Byte) |
int32_t vqdmulhs_lane_s32 (int32_t a, int32x2_t v, const int lane) A64: SQDMULH Sd, Sn, Vm.S[lane] |
MultiplyDoublingWideningAndAddSaturateScalar(Vector64<Int32>, Vector64<Int16>, Vector64<Int16>) |
int32_t vqdmlalh_s16 (int32_t a, int16_t b, int16_t c) A64: SQDMLAL Sd, Hn, Hm |
MultiplyDoublingWideningAndAddSaturateScalar(Vector64<Int64>, Vector64<Int32>, Vector64<Int32>) |
int64_t vqdmlals_s32 (int64_t a, int32_t b, int32_t c) A64: SQDMLAL Dd, Sn, Sm |
MultiplyDoublingWideningAndSubtractSaturateScalar(Vector64<Int32>, Vector64<Int16>, Vector64<Int16>) |
int32_t vqdmlslh_s16 (int32_t a, int16_t b, int16_t c) A64: SQDMLSL Sd, Hn, Hm |
MultiplyDoublingWideningAndSubtractSaturateScalar(Vector64<Int64>, Vector64<Int32>, Vector64<Int32>) |
int64_t vqdmlsls_s32 (int64_t a, int32_t b, int32_t c) A64: SQDMLSL Dd, Sn, Sm |
MultiplyDoublingWideningSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
int32_t vqdmullh_s16 (int16_t a, int16_t b) A64: SQDMULL Sd, Hn, Hm |
MultiplyDoublingWideningSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
int64_t vqdmulls_s32 (int32_t a, int32_t b) A64: SQDMULL Dd, Sn, Sm |
MultiplyDoublingWideningSaturateScalarBySelectedScalar(Vector64<Int16>, Vector128<Int16>, Byte) |
int32_t vqdmullh_laneq_s16 (int16_t a, int16x8_t v, const int lane) A64: SQDMULL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningSaturateScalarBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Byte) |
int32_t vqdmullh_lane_s16 (int16_t a, int16x4_t v, const int lane) A64: SQDMULL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningSaturateScalarBySelectedScalar(Vector64<Int32>, Vector128<Int32>, Byte) |
int64_t vqdmulls_laneq_s32 (int32_t a, int32x4_t v, const int lane) A64: SQDMULL Dd, Sn, Vm.S[lane] |
MultiplyDoublingWideningSaturateScalarBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Byte) |
int64_t vqdmulls_lane_s32 (int32_t a, int32x2_t v, const int lane) A64: SQDMULL Dd, Sn, Vm.S[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndAddSaturate(Vector64<Int32>, Vector64<Int16>, Vector128<Int16>, Byte) |
int32_t vqdmlalh_laneq_s16 (int32_t a, int16_t b, int16x8_t v, const int lane) A64: SQDMLAL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndAddSaturate(Vector64<Int32>, Vector64<Int16>, Vector64<Int16>, Byte) |
int32_t vqdmlalh_lane_s16 (int32_t a, int16_t b, int16x4_t v, const int lane) A64: SQDMLAL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndAddSaturate(Vector64<Int64>, Vector64<Int32>, Vector128<Int32>, Byte) |
int64_t vqdmlals_laneq_s32 (int64_t a, int32_t b, int32x4_t v, const int lane) A64: SQDMLAL Dd, Sn, Vm.S[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndAddSaturate(Vector64<Int64>, Vector64<Int32>, Vector64<Int32>, Byte) |
int64_t vqdmlals_lane_s32 (int64_t a, int32_t b, int32x2_t v, const int lane) A64: SQDMLAL Dd, Sn, Vm.S[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndSubtractSaturate(Vector64<Int32>, Vector64<Int16>, Vector128<Int16>, Byte) |
int32_t vqdmlslh_laneq_s16 (int32_t a, int16_t b, int16x8_t v, const int lane) A64: SQDMLSL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndSubtractSaturate(Vector64<Int32>, Vector64<Int16>, Vector64<Int16>, Byte) |
int32_t vqdmlslh_lane_s16 (int32_t a, int16_t b, int16x4_t v, const int lane) A64: SQDMLSL Sd, Hn, Vm.H[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndSubtractSaturate(Vector64<Int64>, Vector64<Int32>, Vector128<Int32>, Byte) |
int64_t vqdmlsls_laneq_s32 (int64_t a, int32_t b, int32x4_t v, const int lane) A64: SQDMLSL Dd, Sn, Vm.S[lane] |
MultiplyDoublingWideningScalarBySelectedScalarAndSubtractSaturate(Vector64<Int64>, Vector64<Int32>, Vector64<Int32>, Byte) |
int64_t vqdmlsls_lane_s32 (int64_t a, int32_t b, int32x2_t v, const int lane) A64: SQDMLSL Dd, Sn, Vm.S[lane] |
MultiplyExtended(Vector128<Double>, Vector128<Double>) |
float64x2_t vmulxq_f64 (float64x2_t a, float64x2_t b) A64: FMULX Vd.2D, Vn.2D, Vm.2D |
MultiplyExtended(Vector128<Single>, Vector128<Single>) |
float32x4_t vmulxq_f32 (float32x4_t a, float32x4_t b) A64: FMULX Vd.4S, Vn.4S, Vm.4S |
MultiplyExtended(Vector64<Single>, Vector64<Single>) |
float32x2_t vmulx_f32 (float32x2_t a, float32x2_t b) A64: FMULX Vd.2S, Vn.2S, Vm.2S |
MultiplyExtendedByScalar(Vector128<Double>, Vector64<Double>) |
float64x2_t vmulxq_lane_f64 (float64x2_t a, float64x1_t v, const int lane) A64: FMULX Vd.2D, Vn.2D, Vm.D[0] |
MultiplyExtendedBySelectedScalar(Vector128<Double>, Vector128<Double>, Byte) |
float64x2_t vmulxq_laneq_f64 (float64x2_t a, float64x2_t v, const int lane) A64: FMULX Vd.2D, Vn.2D, Vm.D[lane] |
MultiplyExtendedBySelectedScalar(Vector128<Single>, Vector128<Single>, Byte) |
float32x4_t vmulxq_laneq_f32 (float32x4_t a, float32x4_t v, const int lane) A64: FMULX Vd.4S, Vn.4S, Vm.S[lane] |
MultiplyExtendedBySelectedScalar(Vector128<Single>, Vector64<Single>, Byte) |
float32x4_t vmulxq_lane_f32 (float32x4_t a, float32x2_t v, const int lane) A64: FMULX Vd.4S, Vn.4S, Vm.S[lane] |
MultiplyExtendedBySelectedScalar(Vector64<Single>, Vector128<Single>, Byte) |
float32x2_t vmulx_laneq_f32 (float32x2_t a, float32x4_t v, const int lane) A64: FMULX Vd.2S, Vn.2S, Vm.S[lane] |
MultiplyExtendedBySelectedScalar(Vector64<Single>, Vector64<Single>, Byte) |
float32x2_t vmulx_lane_f32 (float32x2_t a, float32x2_t v, const int lane) A64: FMULX Vd.2S, Vn.2S, Vm.S[lane] |
MultiplyExtendedScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vmulx_f64 (float64x1_t a, float64x1_t b) A64: FMULX Dd, Dn, Dm |
MultiplyExtendedScalar(Vector64<Single>, Vector64<Single>) |
float32_t vmulxs_f32 (float32_t a, float32_t b) A64: FMULX Sd, Sn, Sm |
MultiplyExtendedScalarBySelectedScalar(Vector64<Double>, Vector128<Double>, Byte) |
float64_t vmulxd_laneq_f64 (float64_t a, float64x2_t v, const int lane) A64: FMULX Dd, Dn, Vm.D[lane] |
MultiplyExtendedScalarBySelectedScalar(Vector64<Single>, Vector128<Single>, Byte) |
float32_t vmulxs_laneq_f32 (float32_t a, float32x4_t v, const int lane) A64: FMULX Sd, Sn, Vm.S[lane] |
MultiplyExtendedScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Byte) |
float32_t vmulxs_lane_f32 (float32_t a, float32x2_t v, const int lane) A64: FMULX Sd, Sn, Vm.S[lane] |
MultiplyRoundedDoublingSaturateHighScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqrdmulhh_s16 (int16_t a, int16_t b) A64: SQRDMULH Hd, Hn, Hm |
MultiplyRoundedDoublingSaturateHighScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqrdmulhs_s32 (int32_t a, int32_t b) A64: SQRDMULH Sd, Sn, Sm |
MultiplyRoundedDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int16>, Vector128<Int16>, Byte) |
int16_t vqrdmulhh_laneq_s16 (int16_t a, int16x8_t v, const int lane) A64: SQRDMULH Hd, Hn, Vm.H[lane] |
MultiplyRoundedDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>, Byte) |
int16_t vqrdmulhh_lane_s16 (int16_t a, int16x4_t v, const int lane) A64: SQRDMULH Hd, Hn, Vm.H[lane] |
MultiplyRoundedDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int32>, Vector128<Int32>, Byte) |
int32_t vqrdmulhs_laneq_s32 (int32_t a, int32x4_t v, const int lane) A64: SQRDMULH Sd, Sn, Vm.S[lane] |
MultiplyRoundedDoublingScalarBySelectedScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>, Byte) |
int32_t vqrdmulhs_lane_s32 (int32_t a, int32x2_t v, const int lane) A64: SQRDMULH Sd, Sn, Vm.S[lane] |
MultiplyScalarBySelectedScalar(Vector64<Double>, Vector128<Double>, Byte) |
float64_t vmuld_laneq_f64 (float64_t a, float64x2_t v, const int lane) A64: FMUL Dd, Dn, Vm.D[lane] |
Negate(Vector128<Double>) |
float64x2_t vnegq_f64 (float64x2_t a) A64: FNEG Vd.2D, Vn.2D |
Negate(Vector128<Int64>) |
int64x2_t vnegq_s64 (int64x2_t a) A64: NEG Vd.2D, Vn.2D |
NegateSaturate(Vector128<Int64>) |
int64x2_t vqnegq_s64 (int64x2_t a) A64: SQNEG Vd.2D, Vn.2D |
NegateSaturateScalar(Vector64<Int16>) |
int16_t vqnegh_s16 (int16_t a) A64: SQNEG Hd, Hn |
NegateSaturateScalar(Vector64<Int32>) |
int32_t vqnegs_s32 (int32_t a) A64: SQNEG Sd, Sn |
NegateSaturateScalar(Vector64<Int64>) |
int64_t vqnegd_s64 (int64_t a) A64: SQNEG Dd, Dn |
NegateSaturateScalar(Vector64<SByte>) |
int8_t vqnegb_s8 (int8_t a) A64: SQNEG Bd, Bn |
NegateScalar(Vector64<Int64>) |
int64x1_t vneg_s64 (int64x1_t a) A64: NEG Dd, Dn |
ReciprocalEstimate(Vector128<Double>) |
float64x2_t vrecpeq_f64 (float64x2_t a) A64: FRECPE Vd.2D, Vn.2D |
ReciprocalEstimateScalar(Vector64<Double>) |
float64x1_t vrecpe_f64 (float64x1_t a) A64: FRECPE Dd, Dn |
ReciprocalEstimateScalar(Vector64<Single>) |
float32_t vrecpes_f32 (float32_t a) A64: FRECPE Sd, Sn |
ReciprocalExponentScalar(Vector64<Double>) |
float64_t vrecpxd_f64 (float64_t a) A64: FRECPX Dd, Dn |
ReciprocalExponentScalar(Vector64<Single>) |
float32_t vrecpxs_f32 (float32_t a) A64: FRECPX Sd, Sn |
ReciprocalSquareRootEstimate(Vector128<Double>) |
float64x2_t vrsqrteq_f64 (float64x2_t a) A64: FRSQRTE Vd.2D, Vn.2D |
ReciprocalSquareRootEstimateScalar(Vector64<Double>) |
float64x1_t vrsqrte_f64 (float64x1_t a) A64: FRSQRTE Dd, Dn |
ReciprocalSquareRootEstimateScalar(Vector64<Single>) |
float32_t vrsqrtes_f32 (float32_t a) A64: FRSQRTE Sd, Sn |
ReciprocalSquareRootStep(Vector128<Double>, Vector128<Double>) |
float64x2_t vrsqrtsq_f64 (float64x2_t a, float64x2_t b) A64: FRSQRTS Vd.2D, Vn.2D, Vm.2D |
ReciprocalSquareRootStepScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vrsqrts_f64 (float64x1_t a, float64x1_t b) A64: FRSQRTS Dd, Dn, Dm |
ReciprocalSquareRootStepScalar(Vector64<Single>, Vector64<Single>) |
float32_t vrsqrtss_f32 (float32_t a, float32_t b) A64: FRSQRTS Sd, Sn, Sm |
ReciprocalStep(Vector128<Double>, Vector128<Double>) |
float64x2_t vrecpsq_f64 (float64x2_t a, float64x2_t b) A64: FRECPS Vd.2D, Vn.2D, Vm.2D |
ReciprocalStepScalar(Vector64<Double>, Vector64<Double>) |
float64x1_t vrecps_f64 (float64x1_t a, float64x1_t b) A64: FRECPS Dd, Dn, Dm |
ReciprocalStepScalar(Vector64<Single>, Vector64<Single>) |
float32_t vrecpss_f32 (float32_t a, float32_t b) A64: FRECPS Sd, Sn, Sm |
ReverseElementBits(Vector128<Byte>) |
uint8x16_t vrbitq_u8 (uint8x16_t a) A64: RBIT Vd.16B, Vn.16B |
ReverseElementBits(Vector128<SByte>) |
int8x16_t vrbitq_s8 (int8x16_t a) A64: RBIT Vd.16B, Vn.16B |
ReverseElementBits(Vector64<Byte>) |
uint8x8_t vrbit_u8 (uint8x8_t a) A64: RBIT Vd.8B, Vn.8B |
ReverseElementBits(Vector64<SByte>) |
int8x8_t vrbit_s8 (int8x8_t a) A64: RBIT Vd.8B, Vn.8B |
RoundAwayFromZero(Vector128<Double>) |
float64x2_t vrndaq_f64 (float64x2_t a) A64: FRINTA Vd.2D, Vn.2D |
RoundToNearest(Vector128<Double>) |
float64x2_t vrndnq_f64 (float64x2_t a) A64: FRINTN Vd.2D, Vn.2D |
RoundToNegativeInfinity(Vector128<Double>) |
float64x2_t vrndmq_f64 (float64x2_t a) A64: FRINTM Vd.2D, Vn.2D |
RoundToPositiveInfinity(Vector128<Double>) |
float64x2_t vrndpq_f64 (float64x2_t a) A64: FRINTP Vd.2D, Vn.2D |
RoundToZero(Vector128<Double>) |
float64x2_t vrndq_f64 (float64x2_t a) A64: FRINTZ Vd.2D, Vn.2D |
ShiftArithmeticRoundedSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqrshlh_s16 (int16_t a, int16_t b) A64: SQRSHL Hd, Hn, Hm |
ShiftArithmeticRoundedSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqrshls_s32 (int32_t a, int32_t b) A64: SQRSHL Sd, Sn, Sm |
ShiftArithmeticRoundedSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
int8_t vqrshlb_s8 (int8_t a, int8_t b) A64: SQRSHL Bd, Bn, Bm |
ShiftArithmeticSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqshlh_s16 (int16_t a, int16_t b) A64: SQSHL Hd, Hn, Hm |
ShiftArithmeticSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqshls_s32 (int32_t a, int32_t b) A64: SQSHL Sd, Sn, Sm |
ShiftArithmeticSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
int8_t vqshlb_s8 (int8_t a, int8_t b) A64: SQSHL Bd, Bn, Bm |
ShiftLeftLogicalSaturateScalar(Vector64<Byte>, Byte) |
uint8_t vqshlb_n_u8 (uint8_t a, const int n) A64: UQSHL Bd, Bn, #n |
ShiftLeftLogicalSaturateScalar(Vector64<Int16>, Byte) |
int16_t vqshlh_n_s16 (int16_t a, const int n) A64: SQSHL Hd, Hn, #n |
ShiftLeftLogicalSaturateScalar(Vector64<Int32>, Byte) |
int32_t vqshls_n_s32 (int32_t a, const int n) A64: SQSHL Sd, Sn, #n |
ShiftLeftLogicalSaturateScalar(Vector64<SByte>, Byte) |
int8_t vqshlb_n_s8 (int8_t a, const int n) A64: SQSHL Bd, Bn, #n |
ShiftLeftLogicalSaturateScalar(Vector64<UInt16>, Byte) |
uint16_t vqshlh_n_u16 (uint16_t a, const int n) A64: UQSHL Hd, Hn, #n |
ShiftLeftLogicalSaturateScalar(Vector64<UInt32>, Byte) |
uint32_t vqshls_n_u32 (uint32_t a, const int n) A64: UQSHL Sd, Sn, #n |
ShiftLeftLogicalSaturateUnsignedScalar(Vector64<Int16>, Byte) |
uint16_t vqshluh_n_s16 (int16_t a, const int n) A64: SQSHLU Hd, Hn, #n |
ShiftLeftLogicalSaturateUnsignedScalar(Vector64<Int32>, Byte) |
uint32_t vqshlus_n_s32 (int32_t a, const int n) A64: SQSHLU Sd, Sn, #n |
ShiftLeftLogicalSaturateUnsignedScalar(Vector64<SByte>, Byte) |
uint8_t vqshlub_n_s8 (int8_t a, const int n) A64: SQSHLU Bd, Bn, #n |
ShiftLogicalRoundedSaturateScalar(Vector64<Byte>, Vector64<SByte>) |
uint8_t vqrshlb_u8 (uint8_t a, int8_t b) A64: UQRSHL Bd, Bn, Bm |
ShiftLogicalRoundedSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
uint16_t vqrshlh_u16 (uint16_t a, int16_t b) A64: UQRSHL Hd, Hn, Hm |
ShiftLogicalRoundedSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
uint32_t vqrshls_u32 (uint32_t a, int32_t b) A64: UQRSHL Sd, Sn, Sm |
ShiftLogicalRoundedSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
uint8_t vqrshlb_u8 (uint8_t a, int8_t b) A64: UQRSHL Bd, Bn, Bm |
ShiftLogicalRoundedSaturateScalar(Vector64<UInt16>, Vector64<Int16>) |
uint16_t vqrshlh_u16 (uint16_t a, int16_t b) A64: UQRSHL Hd, Hn, Hm |
ShiftLogicalRoundedSaturateScalar(Vector64<UInt32>, Vector64<Int32>) |
uint32_t vqrshls_u32 (uint32_t a, int32_t b) A64: UQRSHL Sd, Sn, Sm |
ShiftLogicalSaturateScalar(Vector64<Byte>, Vector64<SByte>) |
uint8_t vqshlb_u8 (uint8_t a, int8_t b) A64: UQSHL Bd, Bn, Bm |
ShiftLogicalSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
uint16_t vqshlh_u16 (uint16_t a, int16_t b) A64: UQSHL Hd, Hn, Hm |
ShiftLogicalSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
uint32_t vqshls_u32 (uint32_t a, int32_t b) A64: UQSHL Sd, Sn, Sm |
ShiftLogicalSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
uint8_t vqshlb_u8 (uint8_t a, int8_t b) A64: UQSHL Bd, Bn, Bm |
ShiftLogicalSaturateScalar(Vector64<UInt16>, Vector64<Int16>) |
uint16_t vqshlh_u16 (uint16_t a, int16_t b) A64: UQSHL Hd, Hn, Hm |
ShiftLogicalSaturateScalar(Vector64<UInt32>, Vector64<Int32>) |
uint32_t vqshls_u32 (uint32_t a, int32_t b) A64: UQSHL Sd, Sn, Sm |
ShiftRightArithmeticNarrowingSaturateScalar(Vector64<Int16>, Byte) |
int8_t vqshrnh_n_s16 (int16_t a, const int n) A64: SQSHRN Bd, Hn, #n |
ShiftRightArithmeticNarrowingSaturateScalar(Vector64<Int32>, Byte) |
int16_t vqshrns_n_s32 (int32_t a, const int n) A64: SQSHRN Hd, Sn, #n |
ShiftRightArithmeticNarrowingSaturateScalar(Vector64<Int64>, Byte) |
int32_t vqshrnd_n_s64 (int64_t a, const int n) A64: SQSHRN Sd, Dn, #n |
ShiftRightArithmeticNarrowingSaturateUnsignedScalar(Vector64<Int16>, Byte) |
uint8_t vqshrunh_n_s16 (int16_t a, const int n) A64: SQSHRUN Bd, Hn, #n |
ShiftRightArithmeticNarrowingSaturateUnsignedScalar(Vector64<Int32>, Byte) |
uint16_t vqshruns_n_s32 (int32_t a, const int n) A64: SQSHRUN Hd, Sn, #n |
ShiftRightArithmeticNarrowingSaturateUnsignedScalar(Vector64<Int64>, Byte) |
uint32_t vqshrund_n_s64 (int64_t a, const int n) A64: SQSHRUN Sd, Dn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateScalar(Vector64<Int16>, Byte) |
int8_t vqrshrnh_n_s16 (int16_t a, const int n) A64: SQRSHRN Bd, Hn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateScalar(Vector64<Int32>, Byte) |
int16_t vqrshrns_n_s32 (int32_t a, const int n) A64: SQRSHRN Hd, Sn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateScalar(Vector64<Int64>, Byte) |
int32_t vqrshrnd_n_s64 (int64_t a, const int n) A64: SQRSHRN Sd, Dn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateUnsignedScalar(Vector64<Int16>, Byte) |
uint8_t vqrshrunh_n_s16 (int16_t a, const int n) A64: SQRSHRUN Bd, Hn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateUnsignedScalar(Vector64<Int32>, Byte) |
uint16_t vqrshruns_n_s32 (int32_t a, const int n) A64: SQRSHRUN Hd, Sn, #n |
ShiftRightArithmeticRoundedNarrowingSaturateUnsignedScalar(Vector64<Int64>, Byte) |
uint32_t vqrshrund_n_s64 (int64_t a, const int n) A64: SQRSHRUN Sd, Dn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<Int16>, Byte) |
uint8_t vqshrnh_n_u16 (uint16_t a, const int n) A64: UQSHRN Bd, Hn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<Int32>, Byte) |
uint16_t vqshrns_n_u32 (uint32_t a, const int n) A64: UQSHRN Hd, Sn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<Int64>, Byte) |
uint32_t vqshrnd_n_u64 (uint64_t a, const int n) A64: UQSHRN Sd, Dn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<UInt16>, Byte) |
uint8_t vqshrnh_n_u16 (uint16_t a, const int n) A64: UQSHRN Bd, Hn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<UInt32>, Byte) |
uint16_t vqshrns_n_u32 (uint32_t a, const int n) A64: UQSHRN Hd, Sn, #n |
ShiftRightLogicalNarrowingSaturateScalar(Vector64<UInt64>, Byte) |
uint32_t vqshrnd_n_u64 (uint64_t a, const int n) A64: UQSHRN Sd, Dn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<Int16>, Byte) |
Obsolete.
uint8_t vqrshrnh_n_u16 (uint16_t a, const int n) A64: UQRSHRN Bd, Hn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<Int32>, Byte) |
Obsolete.
uint16_t vqrshrns_n_u32 (uint32_t a, const int n) A64: UQRSHRN Hd, Sn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<Int64>, Byte) |
Obsolete.
uint32_t vqrshrnd_n_u64 (uint64_t a, const int n) A64: UQRSHRN Sd, Dn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<UInt16>, Byte) |
uint8_t vqrshrnh_n_u16 (uint16_t a, const int n) A64: UQRSHRN Bd, Hn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<UInt32>, Byte) |
uint16_t vqrshrns_n_u32 (uint32_t a, const int n) A64: UQRSHRN Hd, Sn, #n |
ShiftRightLogicalRoundedNarrowingSaturateScalar(Vector64<UInt64>, Byte) |
uint32_t vqrshrnd_n_u64 (uint64_t a, const int n) A64: UQRSHRN Sd, Dn, #n |
Sqrt(Vector128<Double>) |
float64x2_t vsqrtq_f64 (float64x2_t a) A64: FSQRT Vd.2D, Vn.2D |
Sqrt(Vector128<Single>) |
float32x4_t vsqrtq_f32 (float32x4_t a) A64: FSQRT Vd.4S, Vn.4S |
Sqrt(Vector64<Single>) |
float32x2_t vsqrt_f32 (float32x2_t a) A64: FSQRT Vd.2S, Vn.2S |
Store(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>) |
A64: ST1 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Store(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>) |
A64: ST1 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Store(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>>) |
A64: ST1 { Vn.16B, Vn+1.16B }, [Xn] |
Store(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>,Vector128<Double>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Store(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Store(Double*, ValueTuple<Vector128<Double>,Vector128<Double>>) |
A64: ST1 { Vn.2D, Vn+1.2D }, [Xn] |
Store(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>) |
A64: ST1 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Store(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>) |
A64: ST1 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Store(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>>) |
A64: ST1 { Vn.8H, Vn+1.8H }, [Xn] |
Store(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Store(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Store(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>>) |
A64: ST1 { Vn.4S, Vn+1.4S }, [Xn] |
Store(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Store(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Store(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>>) |
A64: ST1 { Vn.2D, Vn+1.2D }, [Xn] |
Store(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>) |
A64: ST1 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
Store(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>) |
A64: ST1 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
Store(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>>) |
A64: ST1 { Vn.16B, Vn+1.16B }, [Xn] |
Store(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>,Vector128<Single>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Store(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Store(Single*, ValueTuple<Vector128<Single>,Vector128<Single>>) |
A64: ST1 { Vn.4S, Vn+1.4S }, [Xn] |
Store(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST1 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
Store(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST1 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
Store(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST1 { Vn.8H, Vn+1.8H }, [Xn] |
Store(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
Store(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST1 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
Store(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST1 { Vn.4S, Vn+1.4S }, [Xn] |
Store(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
Store(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST1 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
Store(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST1 { Vn.2D, Vn+1.2D }, [Xn] |
StorePair(Byte*, Vector128<Byte>, Vector128<Byte>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Byte*, Vector64<Byte>, Vector64<Byte>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(Double*, Vector128<Double>, Vector128<Double>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Double*, Vector64<Double>, Vector64<Double>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(Int16*, Vector128<Int16>, Vector128<Int16>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Int16*, Vector64<Int16>, Vector64<Int16>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(Int32*, Vector128<Int32>, Vector128<Int32>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Int32*, Vector64<Int32>, Vector64<Int32>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(Int64*, Vector128<Int64>, Vector128<Int64>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Int64*, Vector64<Int64>, Vector64<Int64>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(SByte*, Vector128<SByte>, Vector128<SByte>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(SByte*, Vector64<SByte>, Vector64<SByte>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(Single*, Vector128<Single>, Vector128<Single>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(Single*, Vector64<Single>, Vector64<Single>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(UInt16*, Vector128<UInt16>, Vector128<UInt16>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(UInt16*, Vector64<UInt16>, Vector64<UInt16>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(UInt32*, Vector128<UInt32>, Vector128<UInt32>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(UInt32*, Vector64<UInt32>, Vector64<UInt32>) |
A64: STP Dt1, Dt2, [Xn] |
StorePair(UInt64*, Vector128<UInt64>, Vector128<UInt64>) |
A64: STP Qt1, Qt2, [Xn] |
StorePair(UInt64*, Vector64<UInt64>, Vector64<UInt64>) |
A64: STP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Byte*, Vector128<Byte>, Vector128<Byte>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Byte*, Vector64<Byte>, Vector64<Byte>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Double*, Vector128<Double>, Vector128<Double>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Double*, Vector64<Double>, Vector64<Double>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Int16*, Vector128<Int16>, Vector128<Int16>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Int16*, Vector64<Int16>, Vector64<Int16>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Int32*, Vector128<Int32>, Vector128<Int32>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Int32*, Vector64<Int32>, Vector64<Int32>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Int64*, Vector128<Int64>, Vector128<Int64>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Int64*, Vector64<Int64>, Vector64<Int64>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(SByte*, Vector128<SByte>, Vector128<SByte>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(SByte*, Vector64<SByte>, Vector64<SByte>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(Single*, Vector128<Single>, Vector128<Single>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(Single*, Vector64<Single>, Vector64<Single>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(UInt16*, Vector128<UInt16>, Vector128<UInt16>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(UInt16*, Vector64<UInt16>, Vector64<UInt16>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(UInt32*, Vector128<UInt32>, Vector128<UInt32>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(UInt32*, Vector64<UInt32>, Vector64<UInt32>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairNonTemporal(UInt64*, Vector128<UInt64>, Vector128<UInt64>) |
A64: STNP Qt1, Qt2, [Xn] |
StorePairNonTemporal(UInt64*, Vector64<UInt64>, Vector64<UInt64>) |
A64: STNP Dt1, Dt2, [Xn] |
StorePairScalar(Int32*, Vector64<Int32>, Vector64<Int32>) |
A64: STP St1, St2, [Xn] |
StorePairScalar(Single*, Vector64<Single>, Vector64<Single>) |
A64: STP St1, St2, [Xn] |
StorePairScalar(UInt32*, Vector64<UInt32>, Vector64<UInt32>) |
A64: STP St1, St2, [Xn] |
StorePairScalarNonTemporal(Int32*, Vector64<Int32>, Vector64<Int32>) |
A64: STNP St1, St2, [Xn] |
StorePairScalarNonTemporal(Single*, Vector64<Single>, Vector64<Single>) |
A64: STNP St1, St2, [Xn] |
StorePairScalarNonTemporal(UInt32*, Vector64<UInt32>, Vector64<UInt32>) |
A64: STNP St1, St2, [Xn] |
StoreSelectedScalar(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Byte) |
void vst4_lane_s8 (int8_t * ptr, int8x16x4_t val, const int lane) A64: ST4 { Vt.16B, Vt+1.16B, Vt+2.16B, Vt+3.16B }[index], [Xn] |
StoreSelectedScalar(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Byte) |
void vst3_lane_s8 (int8_t * ptr, int8x16x3_t val, const int lane) A64: ST3 { Vt.16B, Vt+1.16B, Vt+2.16B }[index], [Xn] |
StoreSelectedScalar(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>>, Byte) |
void vst2_lane_s8 (int8_t * ptr, int8x16x2_t val, const int lane) A64: ST2 { Vt.16B, Vt+1.16B }[index], [Xn] |
StoreSelectedScalar(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>,Vector128<Double>>, Byte) |
A64: ST4 { Vt.2D, Vt+1.2D, Vt+2.2D, Vt+3.2D }[index], [Xn] |
StoreSelectedScalar(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>>, Byte) |
A64: ST3 { Vt.2D, Vt+1.2D, Vt+2.2D }[index], [Xn] |
StoreSelectedScalar(Double*, ValueTuple<Vector128<Double>,Vector128<Double>>, Byte) |
A64: ST2 { Vt.2D, Vt+1.2D }[index], [Xn] |
StoreSelectedScalar(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>, Byte) |
void vst4_lane_s16 (int16_t * ptr, int16x8x4_t val, const int lane) A64: ST4 { Vt.8H, Vt+1.8H, Vt+2.8H, Vt+3.8H }[index], [Xn] |
StoreSelectedScalar(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>, Byte) |
void vst3_lane_s16 (int16_t * ptr, int16x8x3_t val, const int lane) A64: ST3 { Vt.8H, Vt+1.8H, Vt+2.8H }[index], [Xn] |
StoreSelectedScalar(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>>, Byte) |
void vst2_lane_s16 (int16_t * ptr, int16x8x2_t val, const int lane) A64: ST2 { Vt.8H, Vt+1.8H }[index], [Xn] |
StoreSelectedScalar(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>, Byte) |
void vst4_lane_s32 (int32_t * ptr, int32x4x4_t val, const int lane) A64: ST4 { Vt.4S, Vt+1.4S, Vt+2.4S, Vt+3.4S }[index], [Xn] |
StoreSelectedScalar(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>, Byte) |
void vst3_lane_s32 (int32_t * ptr, int32x4x3_t val, const int lane) A64: ST3 { Vt.4S, Vt+1.4S, Vt+2.4S }[index], [Xn] |
StoreSelectedScalar(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>>, Byte) |
void vst2_lane_s32 (int32_t * ptr, int32x4x2_t val, const int lane) A64: ST2 { Vt.4S, Vt+1.4S }[index], [Xn] |
StoreSelectedScalar(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>, Byte) |
A64: ST4 { Vt.2D, Vt+1.2D, Vt+2.2D, Vt+3.2D }[index], [Xn] |
StoreSelectedScalar(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>, Byte) |
A64: ST3 { Vt.2D, Vt+1.2D, Vt+2.2D }[index], [Xn] |
StoreSelectedScalar(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>>, Byte) |
A64: ST2 { Vt.2D, Vt+1.2D }[index], [Xn] |
StoreSelectedScalar(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Byte) |
void vst4_lane_s8 (int8_t * ptr, int8x16x4_t val, const int lane) A64: ST4 { Vt.16B, Vt+1.16B, Vt+2.16B, Vt+3.16B }[index], [Xn] |
StoreSelectedScalar(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Byte) |
void vst3_lane_s8 (int8_t * ptr, int8x16x3_t val, const int lane) A64: ST3 { Vt.16B, Vt+1.16B, Vt+2.16B }[index], [Xn] |
StoreSelectedScalar(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>>, Byte) |
void vst2_lane_s8 (int8_t * ptr, int8x16x2_t val, const int lane) A64: ST2 { Vt.16B, Vt+1.16B }[index], [Xn] |
StoreSelectedScalar(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>,Vector128<Single>>, Byte) |
void vst4_lane_f32 (float32_t * ptr, float32x2x4_t val, const int lane) A64: ST4 { Vt.4S, Vt+1.4S, Vt+2.4S, Vt+3.4S }[index], [Xn] |
StoreSelectedScalar(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>>, Byte) |
void vst3_lane_f32 (float32_t * ptr, float32x2x3_t val, const int lane) A64: ST3 { Vt.4S, Vt+1.4S, Vt+2.4S }[index], [Xn] |
StoreSelectedScalar(Single*, ValueTuple<Vector128<Single>,Vector128<Single>>, Byte) |
void vst2_lane_f32 (float32_t * ptr, float32x2x2_t val, const int lane) A64: ST2 { Vt.4S, Vt+1.4S }[index], [Xn] |
StoreSelectedScalar(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>, Byte) |
void vst4_lane_s16 (int16_t * ptr, int16x8x4_t val, const int lane) A64: ST4 { Vt.8H, Vt+1.8H, Vt+2.8H, Vt+3.8H }[index], [Xn] |
StoreSelectedScalar(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>, Byte) |
void vst3_lane_s16 (int16_t * ptr, int16x8x3_t val, const int lane) A64: ST3 { Vt.8H, Vt+1.8H, Vt+2.8H }[index], [Xn] |
StoreSelectedScalar(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>>, Byte) |
void vst2_lane_s16 (int16_t * ptr, int16x8x2_t val, const int lane) A64: ST2 { Vt.8H, Vt+1.8H }[index], [Xn] |
StoreSelectedScalar(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>, Byte) |
void vst4_lane_s32 (int32_t * ptr, int32x4x4_t val, const int lane) A64: ST4 { Vt.4S, Vt+1.4S, Vt+2.4S, Vt+3.4S }[index], [Xn] |
StoreSelectedScalar(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>, Byte) |
void vst3_lane_s32 (int32_t * ptr, int32x4x3_t val, const int lane) A64: ST3 { Vt.4S, Vt+1.4S, Vt+2.4S }[index], [Xn] |
StoreSelectedScalar(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>>, Byte) |
void vst2_lane_s32 (int32_t * ptr, int32x4x2_t val, const int lane) A64: ST2 { Vt.4S, Vt+1.4S }[index], [Xn] |
StoreSelectedScalar(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>, Byte) |
A64: ST4 { Vt.2D, Vt+1.2D, Vt+2.2D, Vt+3.2D }[index], [Xn] |
StoreSelectedScalar(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>, Byte) |
A64: ST3 { Vt.2D, Vt+1.2D, Vt+2.2D }[index], [Xn] |
StoreSelectedScalar(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>>, Byte) |
A64: ST2 { Vt.2D, Vt+1.2D }[index], [Xn] |
StoreVectorAndZip(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>) |
A64: ST4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
StoreVectorAndZip(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>) |
A64: ST3 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
StoreVectorAndZip(Byte*, ValueTuple<Vector128<Byte>,Vector128<Byte>>) |
A64: ST2 { Vn.16B, Vn+1.16B }, [Xn] |
StoreVectorAndZip(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>,Vector128<Double>>) |
A64: ST4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
StoreVectorAndZip(Double*, ValueTuple<Vector128<Double>,Vector128<Double>,Vector128<Double>>) |
A64: ST3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
StoreVectorAndZip(Double*, ValueTuple<Vector128<Double>,Vector128<Double>>) |
A64: ST2 { Vn.2D, Vn+1.2D }, [Xn] |
StoreVectorAndZip(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>) |
A64: ST4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
StoreVectorAndZip(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>,Vector128<Int16>>) |
A64: ST3 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
StoreVectorAndZip(Int16*, ValueTuple<Vector128<Int16>,Vector128<Int16>>) |
A64: ST2 { Vn.8H, Vn+1.8H }, [Xn] |
StoreVectorAndZip(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>) |
A64: ST4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
StoreVectorAndZip(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>,Vector128<Int32>>) |
A64: ST3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
StoreVectorAndZip(Int32*, ValueTuple<Vector128<Int32>,Vector128<Int32>>) |
A64: ST2 { Vn.4S, Vn+1.4S }, [Xn] |
StoreVectorAndZip(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>) |
A64: ST4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
StoreVectorAndZip(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>,Vector128<Int64>>) |
A64: ST3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
StoreVectorAndZip(Int64*, ValueTuple<Vector128<Int64>,Vector128<Int64>>) |
A64: ST2 { Vn.2D, Vn+1.2D }, [Xn] |
StoreVectorAndZip(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>) |
A64: ST4 { Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B }, [Xn] |
StoreVectorAndZip(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>) |
A64: ST3 { Vn.16B, Vn+1.16B, Vn+2.16B }, [Xn] |
StoreVectorAndZip(SByte*, ValueTuple<Vector128<SByte>,Vector128<SByte>>) |
A64: ST2 { Vn.16B, Vn+1.16B }, [Xn] |
StoreVectorAndZip(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>,Vector128<Single>>) |
A64: ST4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
StoreVectorAndZip(Single*, ValueTuple<Vector128<Single>,Vector128<Single>,Vector128<Single>>) |
A64: ST3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
StoreVectorAndZip(Single*, ValueTuple<Vector128<Single>,Vector128<Single>>) |
A64: ST2 { Vn.4S, Vn+1.4S }, [Xn] |
StoreVectorAndZip(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST4 { Vn.8H, Vn+1.8H, Vn+2.8H, Vn+3.8H }, [Xn] |
StoreVectorAndZip(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST3 { Vn.8H, Vn+1.8H, Vn+2.8H }, [Xn] |
StoreVectorAndZip(UInt16*, ValueTuple<Vector128<UInt16>,Vector128<UInt16>>) |
A64: ST2 { Vn.8H, Vn+1.8H }, [Xn] |
StoreVectorAndZip(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST4 { Vn.4S, Vn+1.4S, Vn+2.4S, Vn+3.4S }, [Xn] |
StoreVectorAndZip(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST3 { Vn.4S, Vn+1.4S, Vn+2.4S }, [Xn] |
StoreVectorAndZip(UInt32*, ValueTuple<Vector128<UInt32>,Vector128<UInt32>>) |
A64: ST2 { Vn.4S, Vn+1.4S }, [Xn] |
StoreVectorAndZip(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST4 { Vn.2D, Vn+1.2D, Vn+2.2D, Vn+3.2D }, [Xn] |
StoreVectorAndZip(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST3 { Vn.2D, Vn+1.2D, Vn+2.2D }, [Xn] |
StoreVectorAndZip(UInt64*, ValueTuple<Vector128<UInt64>,Vector128<UInt64>>) |
A64: ST2 { Vn.2D, Vn+1.2D }, [Xn] |
Subtract(Vector128<Double>, Vector128<Double>) |
float64x2_t vsubq_f64 (float64x2_t a, float64x2_t b) A64: FSUB Vd.2D, Vn.2D, Vm.2D |
SubtractSaturateScalar(Vector64<Byte>, Vector64<Byte>) |
uint8_t vqsubb_u8 (uint8_t a, uint8_t b) A64: UQSUB Bd, Bn, Bm |
SubtractSaturateScalar(Vector64<Int16>, Vector64<Int16>) |
int16_t vqsubh_s16 (int16_t a, int16_t b) A64: SQSUB Hd, Hn, Hm |
SubtractSaturateScalar(Vector64<Int32>, Vector64<Int32>) |
int32_t vqsubs_s32 (int32_t a, int32_t b) A64: SQSUB Sd, Sn, Sm |
SubtractSaturateScalar(Vector64<SByte>, Vector64<SByte>) |
int8_t vqsubb_s8 (int8_t a, int8_t b) A64: SQSUB Bd, Bn, Bm |
SubtractSaturateScalar(Vector64<UInt16>, Vector64<UInt16>) |
uint16_t vqsubh_u16 (uint16_t a, uint16_t b) A64: UQSUB Hd, Hn, Hm |
SubtractSaturateScalar(Vector64<UInt32>, Vector64<UInt32>) |
uint32_t vqsubs_u32 (uint32_t a, uint32_t b) A64: UQSUB Sd, Sn, Sm |
ToString() |
Returns a string that represents the current object. (Inherited from Object) |
TransposeEven(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vtrn1q_u8(uint8x16_t a, uint8x16_t b) A64: TRN1 Vd.16B, Vn.16B, Vm.16B |
TransposeEven(Vector128<Double>, Vector128<Double>) |
float64x2_t vtrn1q_f64(float64x2_t a, float64x2_t b) A64: TRN1 Vd.2D, Vn.2D, Vm.2D |
TransposeEven(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vtrn1q_s16(int16x8_t a, int16x8_t b) A64: TRN1 Vd.8H, Vn.8H, Vm.8H |
TransposeEven(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vtrn1q_s32(int32x4_t a, int32x4_t b) A64: TRN1 Vd.4S, Vn.4S, Vm.4S |
TransposeEven(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vtrn1q_s64(int64x2_t a, int64x2_t b) A64: TRN1 Vd.2D, Vn.2D, Vm.2D |
TransposeEven(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vtrn1q_u8(int8x16_t a, int8x16_t b) A64: TRN1 Vd.16B, Vn.16B, Vm.16B |
TransposeEven(Vector128<Single>, Vector128<Single>) |
float32x4_t vtrn1q_f32(float32x4_t a, float32x4_t b) A64: TRN1 Vd.4S, Vn.4S, Vm.4S |
TransposeEven(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vtrn1q_u16(uint16x8_t a, uint16x8_t b) A64: TRN1 Vd.8H, Vn.8H, Vm.8H |
TransposeEven(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vtrn1q_u32(uint32x4_t a, uint32x4_t b) A64: TRN1 Vd.4S, Vn.4S, Vm.4S |
TransposeEven(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vtrn1q_u64(uint64x2_t a, uint64x2_t b) A64: TRN1 Vd.2D, Vn.2D, Vm.2D |
TransposeEven(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vtrn1_u8(uint8x8_t a, uint8x8_t b) A64: TRN1 Vd.8B, Vn.8B, Vm.8B |
TransposeEven(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vtrn1_s16(int16x4_t a, int16x4_t b) A64: TRN1 Vd.4H, Vn.4H, Vm.4H |
TransposeEven(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vtrn1_s32(int32x2_t a, int32x2_t b) A64: TRN1 Vd.2S, Vn.2S, Vm.2S |
TransposeEven(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vtrn1_s8(int8x8_t a, int8x8_t b) A64: TRN1 Vd.8B, Vn.8B, Vm.8B |
TransposeEven(Vector64<Single>, Vector64<Single>) |
float32x2_t vtrn1_f32(float32x2_t a, float32x2_t b) A64: TRN1 Vd.2S, Vn.2S, Vm.2S |
TransposeEven(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vtrn1_u16(uint16x4_t a, uint16x4_t b) A64: TRN1 Vd.4H, Vn.4H, Vm.4H |
TransposeEven(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vtrn1_u32(uint32x2_t a, uint32x2_t b) A64: TRN1 Vd.2S, Vn.2S, Vm.2S |
TransposeOdd(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vtrn2q_u8(uint8x16_t a, uint8x16_t b) A64: TRN2 Vd.16B, Vn.16B, Vm.16B |
TransposeOdd(Vector128<Double>, Vector128<Double>) |
float64x2_t vtrn2q_f64(float64x2_t a, float64x2_t b) A64: TRN2 Vd.2D, Vn.2D, Vm.2D |
TransposeOdd(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vtrn2q_s16(int16x8_t a, int16x8_t b) A64: TRN2 Vd.8H, Vn.8H, Vm.8H |
TransposeOdd(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vtrn2q_s32(int32x4_t a, int32x4_t b) A64: TRN2 Vd.4S, Vn.4S, Vm.4S |
TransposeOdd(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vtrn2q_s64(int64x2_t a, int64x2_t b) A64: TRN2 Vd.2D, Vn.2D, Vm.2D |
TransposeOdd(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vtrn2q_u8(int8x16_t a, int8x16_t b) A64: TRN2 Vd.16B, Vn.16B, Vm.16B |
TransposeOdd(Vector128<Single>, Vector128<Single>) |
float32x4_t vtrn2q_f32(float32x4_t a, float32x4_t b) A64: TRN2 Vd.4S, Vn.4S, Vm.4S |
TransposeOdd(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vtrn2q_u16(uint16x8_t a, uint16x8_t b) A64: TRN2 Vd.8H, Vn.8H, Vm.8H |
TransposeOdd(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vtrn2q_u32(uint32x4_t a, uint32x4_t b) A64: TRN2 Vd.4S, Vn.4S, Vm.4S |
TransposeOdd(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vtrn2q_u64(uint64x2_t a, uint64x2_t b) A64: TRN2 Vd.2D, Vn.2D, Vm.2D |
TransposeOdd(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vtrn2_u8(uint8x8_t a, uint8x8_t b) A64: TRN2 Vd.8B, Vn.8B, Vm.8B |
TransposeOdd(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vtrn2_s16(int16x4_t a, int16x4_t b) A64: TRN2 Vd.4H, Vn.4H, Vm.4H |
TransposeOdd(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vtrn2_s32(int32x2_t a, int32x2_t b) A64: TRN2 Vd.2S, Vn.2S, Vm.2S |
TransposeOdd(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vtrn2_s8(int8x8_t a, int8x8_t b) A64: TRN2 Vd.8B, Vn.8B, Vm.8B |
TransposeOdd(Vector64<Single>, Vector64<Single>) |
float32x2_t vtrn2_f32(float32x2_t a, float32x2_t b) A64: TRN2 Vd.2S, Vn.2S, Vm.2S |
TransposeOdd(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vtrn2_u16(uint16x4_t a, uint16x4_t b) A64: TRN2 Vd.4H, Vn.4H, Vm.4H |
TransposeOdd(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vtrn2_u32(uint32x2_t a, uint32x2_t b) A64: TRN2 Vd.2S, Vn.2S, Vm.2S |
UnzipEven(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vuzp1q_u8(uint8x16_t a, uint8x16_t b) A64: UZP1 Vd.16B, Vn.16B, Vm.16B |
UnzipEven(Vector128<Double>, Vector128<Double>) |
float64x2_t vuzp1q_f64(float64x2_t a, float64x2_t b) A64: UZP1 Vd.2D, Vn.2D, Vm.2D |
UnzipEven(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vuzp1q_s16(int16x8_t a, int16x8_t b) A64: UZP1 Vd.8H, Vn.8H, Vm.8H |
UnzipEven(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vuzp1q_s32(int32x4_t a, int32x4_t b) A64: UZP1 Vd.4S, Vn.4S, Vm.4S |
UnzipEven(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vuzp1q_s64(int64x2_t a, int64x2_t b) A64: UZP1 Vd.2D, Vn.2D, Vm.2D |
UnzipEven(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vuzp1q_u8(int8x16_t a, int8x16_t b) A64: UZP1 Vd.16B, Vn.16B, Vm.16B |
UnzipEven(Vector128<Single>, Vector128<Single>) |
float32x4_t vuzp1q_f32(float32x4_t a, float32x4_t b) A64: UZP1 Vd.4S, Vn.4S, Vm.4S |
UnzipEven(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vuzp1q_u16(uint16x8_t a, uint16x8_t b) A64: UZP1 Vd.8H, Vn.8H, Vm.8H |
UnzipEven(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vuzp1q_u32(uint32x4_t a, uint32x4_t b) A64: UZP1 Vd.4S, Vn.4S, Vm.4S |
UnzipEven(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vuzp1q_u64(uint64x2_t a, uint64x2_t b) A64: UZP1 Vd.2D, Vn.2D, Vm.2D |
UnzipEven(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vuzp1_u8(uint8x8_t a, uint8x8_t b) A64: UZP1 Vd.8B, Vn.8B, Vm.8B |
UnzipEven(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vuzp1_s16(int16x4_t a, int16x4_t b) A64: UZP1 Vd.4H, Vn.4H, Vm.4H |
UnzipEven(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vuzp1_s32(int32x2_t a, int32x2_t b) A64: UZP1 Vd.2S, Vn.2S, Vm.2S |
UnzipEven(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vuzp1_s8(int8x8_t a, int8x8_t b) A64: UZP1 Vd.8B, Vn.8B, Vm.8B |
UnzipEven(Vector64<Single>, Vector64<Single>) |
float32x2_t vuzp1_f32(float32x2_t a, float32x2_t b) A64: UZP1 Vd.2S, Vn.2S, Vm.2S |
UnzipEven(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vuzp1_u16(uint16x4_t a, uint16x4_t b) A64: UZP1 Vd.4H, Vn.4H, Vm.4H |
UnzipEven(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vuzp1_u32(uint32x2_t a, uint32x2_t b) A64: UZP1 Vd.2S, Vn.2S, Vm.2S |
UnzipOdd(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vuzp2q_u8(uint8x16_t a, uint8x16_t b) A64: UZP2 Vd.16B, Vn.16B, Vm.16B |
UnzipOdd(Vector128<Double>, Vector128<Double>) |
float64x2_t vuzp2q_f64(float64x2_t a, float64x2_t b) A64: UZP2 Vd.2D, Vn.2D, Vm.2D |
UnzipOdd(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vuzp2q_s16(int16x8_t a, int16x8_t b) A64: UZP2 Vd.8H, Vn.8H, Vm.8H |
UnzipOdd(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vuzp2q_s32(int32x4_t a, int32x4_t b) A64: UZP2 Vd.4S, Vn.4S, Vm.4S |
UnzipOdd(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vuzp2q_s64(int64x2_t a, int64x2_t b) A64: UZP2 Vd.2D, Vn.2D, Vm.2D |
UnzipOdd(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vuzp2q_u8(int8x16_t a, int8x16_t b) A64: UZP2 Vd.16B, Vn.16B, Vm.16B |
UnzipOdd(Vector128<Single>, Vector128<Single>) |
float32x4_t vuzp2_f32(float32x4_t a, float32x4_t b) A64: UZP2 Vd.4S, Vn.4S, Vm.4S |
UnzipOdd(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vuzp2q_u16(uint16x8_t a, uint16x8_t b) A64: UZP2 Vd.8H, Vn.8H, Vm.8H |
UnzipOdd(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vuzp2q_u32(uint32x4_t a, uint32x4_t b) A64: UZP2 Vd.4S, Vn.4S, Vm.4S |
UnzipOdd(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vuzp2q_u64(uint64x2_t a, uint64x2_t b) A64: UZP2 Vd.2D, Vn.2D, Vm.2D |
UnzipOdd(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vuzp2_u8(uint8x8_t a, uint8x8_t b) A64: UZP2 Vd.8B, Vn.8B, Vm.8B |
UnzipOdd(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vuzp2_s16(int16x4_t a, int16x4_t b) A64: UZP2 Vd.4H, Vn.4H, Vm.4H |
UnzipOdd(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vuzp2_s32(int32x2_t a, int32x2_t b) A64: UZP2 Vd.2S, Vn.2S, Vm.2S |
UnzipOdd(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vuzp2_s8(int8x8_t a, int8x8_t b) A64: UZP2 Vd.8B, Vn.8B, Vm.8B |
UnzipOdd(Vector64<Single>, Vector64<Single>) |
float32x2_t vuzp2_f32(float32x2_t a, float32x2_t b) A64: UZP2 Vd.2S, Vn.2S, Vm.2S |
UnzipOdd(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vuzp2_u16(uint16x4_t a, uint16x4_t b) A64: UZP2 Vd.4H, Vn.4H, Vm.4H |
UnzipOdd(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vuzp2_u32(uint32x2_t a, uint32x2_t b) A64: UZP2 Vd.2S, Vn.2S, Vm.2S |
VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbl4q_u8(uint8x16x4_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.16B |
VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbl3q_u8(uint8x16x3_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.16B |
VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbl2q_u8(uint8x16x2_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B}, Vm.16B |
VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbl4q_s8(int8x16x4_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.16B |
VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbl3q_s8(int8x16x3_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.16B |
VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbl2q_s8(int8x16x2_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B, Vn+1.16B}, Vm.16B |
VectorTableLookup(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vqvtbl1q_u8(uint8x16_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B}, Vm.16B |
VectorTableLookup(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vqvtbl1q_s8(int8x16_t t, uint8x16_t idx) A64: TBL Vd.16B, {Vn.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbx4q_u8(uint8x16x4_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbx3q_u8(uint8x16x3_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>>, Vector128<Byte>) |
uint8x16_t vqtbx2q_u8(uint8x16x2_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<Byte>, Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vqvtbx1q_u8(uint8x16_t r, int8x16_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbx4q_s8(int8x16x4_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbx3q_s8(int8x16x3_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>>, Vector128<SByte>) |
int8x16_t vqtbx2q_s8(int8x16x2_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B, Vn+1.16B}, Vm.16B |
VectorTableLookupExtension(Vector128<SByte>, Vector128<SByte>, Vector128<SByte>) |
int8x16_t vqvtbx1q_s8(int8x16_t r, int8x16_t t, uint8x16_t idx) A64: TBX Vd.16B, {Vn.16B}, Vm.16B |
ZipHigh(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vzip2q_u8(uint8x16_t a, uint8x16_t b) A64: ZIP2 Vd.16B, Vn.16B, Vm.16B |
ZipHigh(Vector128<Double>, Vector128<Double>) |
float64x2_t vzip2q_f64(float64x2_t a, float64x2_t b) A64: ZIP2 Vd.2D, Vn.2D, Vm.2D |
ZipHigh(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vzip2q_s16(int16x8_t a, int16x8_t b) A64: ZIP2 Vd.8H, Vn.8H, Vm.8H |
ZipHigh(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vzip2q_s32(int32x4_t a, int32x4_t b) A64: ZIP2 Vd.4S, Vn.4S, Vm.4S |
ZipHigh(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vzip2q_s64(int64x2_t a, int64x2_t b) A64: ZIP2 Vd.2D, Vn.2D, Vm.2D |
ZipHigh(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vzip2q_u8(int8x16_t a, int8x16_t b) A64: ZIP2 Vd.16B, Vn.16B, Vm.16B |
ZipHigh(Vector128<Single>, Vector128<Single>) |
float32x4_t vzip2q_f32(float32x4_t a, float32x4_t b) A64: ZIP2 Vd.4S, Vn.4S, Vm.4S |
ZipHigh(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vzip2q_u16(uint16x8_t a, uint16x8_t b) A64: ZIP2 Vd.8H, Vn.8H, Vm.8H |
ZipHigh(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vzip2q_u32(uint32x4_t a, uint32x4_t b) A64: ZIP2 Vd.4S, Vn.4S, Vm.4S |
ZipHigh(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vzip2q_u64(uint64x2_t a, uint64x2_t b) A64: ZIP2 Vd.2D, Vn.2D, Vm.2D |
ZipHigh(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vzip2_u8(uint8x8_t a, uint8x8_t b) A64: ZIP2 Vd.8B, Vn.8B, Vm.8B |
ZipHigh(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vzip2_s16(int16x4_t a, int16x4_t b) A64: ZIP2 Vd.4H, Vn.4H, Vm.4H |
ZipHigh(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vzip2_s32(int32x2_t a, int32x2_t b) A64: ZIP2 Vd.2S, Vn.2S, Vm.2S |
ZipHigh(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vzip2_s8(int8x8_t a, int8x8_t b) A64: ZIP2 Vd.8B, Vn.8B, Vm.8B |
ZipHigh(Vector64<Single>, Vector64<Single>) |
float32x2_t vzip2_f32(float32x2_t a, float32x2_t b) A64: ZIP2 Vd.2S, Vn.2S, Vm.2S |
ZipHigh(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vzip2_u16(uint16x4_t a, uint16x4_t b) A64: ZIP2 Vd.4H, Vn.4H, Vm.4H |
ZipHigh(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vzip2_u32(uint32x2_t a, uint32x2_t b) A64: ZIP2 Vd.2S, Vn.2S, Vm.2S |
ZipLow(Vector128<Byte>, Vector128<Byte>) |
uint8x16_t vzip1q_u8(uint8x16_t a, uint8x16_t b) A64: ZIP1 Vd.16B, Vn.16B, Vm.16B |
ZipLow(Vector128<Double>, Vector128<Double>) |
float64x2_t vzip1q_f64(float64x2_t a, float64x2_t b) A64: ZIP1 Vd.2D, Vn.2D, Vm.2D |
ZipLow(Vector128<Int16>, Vector128<Int16>) |
int16x8_t vzip1q_s16(int16x8_t a, int16x8_t b) A64: ZIP1 Vd.8H, Vn.8H, Vm.8H |
ZipLow(Vector128<Int32>, Vector128<Int32>) |
int32x4_t vzip1q_s32(int32x4_t a, int32x4_t b) A64: ZIP1 Vd.4S, Vn.4S, Vm.4S |
ZipLow(Vector128<Int64>, Vector128<Int64>) |
int64x2_t vzip1q_s64(int64x2_t a, int64x2_t b) A64: ZIP1 Vd.2D, Vn.2D, Vm.2D |
ZipLow(Vector128<SByte>, Vector128<SByte>) |
int8x16_t vzip1q_u8(int8x16_t a, int8x16_t b) A64: ZIP1 Vd.16B, Vn.16B, Vm.16B |
ZipLow(Vector128<Single>, Vector128<Single>) |
float32x4_t vzip1q_f32(float32x4_t a, float32x4_t b) A64: ZIP1 Vd.4S, Vn.4S, Vm.4S |
ZipLow(Vector128<UInt16>, Vector128<UInt16>) |
uint16x8_t vzip1q_u16(uint16x8_t a, uint16x8_t b) A64: ZIP1 Vd.8H, Vn.8H, Vm.8H |
ZipLow(Vector128<UInt32>, Vector128<UInt32>) |
uint32x4_t vzip1q_u32(uint32x4_t a, uint32x4_t b) A64: ZIP1 Vd.4S, Vn.4S, Vm.4S |
ZipLow(Vector128<UInt64>, Vector128<UInt64>) |
uint64x2_t vzip1q_u64(uint64x2_t a, uint64x2_t b) A64: ZIP1 Vd.2D, Vn.2D, Vm.2D |
ZipLow(Vector64<Byte>, Vector64<Byte>) |
uint8x8_t vzip1_u8(uint8x8_t a, uint8x8_t b) A64: ZIP1 Vd.8B, Vn.8B, Vm.8B |
ZipLow(Vector64<Int16>, Vector64<Int16>) |
int16x4_t vzip1_s16(int16x4_t a, int16x4_t b) A64: ZIP1 Vd.4H, Vn.4H, Vm.4H |
ZipLow(Vector64<Int32>, Vector64<Int32>) |
int32x2_t vzip1_s32(int32x2_t a, int32x2_t b) A64: ZIP1 Vd.2S, Vn.2S, Vm.2S |
ZipLow(Vector64<SByte>, Vector64<SByte>) |
int8x8_t vzip1_s8(int8x8_t a, int8x8_t b) A64: ZIP1 Vd.8B, Vn.8B, Vm.8B |
ZipLow(Vector64<Single>, Vector64<Single>) |
float32x2_t vzip1_f32(float32x2_t a, float32x2_t b) A64: ZIP1 Vd.2S, Vn.2S, Vm.2S |
ZipLow(Vector64<UInt16>, Vector64<UInt16>) |
uint16x4_t vzip1_u16(uint16x4_t a, uint16x4_t b) A64: ZIP1 Vd.4H, Vn.4H, Vm.4H |
ZipLow(Vector64<UInt32>, Vector64<UInt32>) |
uint32x2_t vzip1_u32(uint32x2_t a, uint32x2_t b) A64: ZIP1 Vd.2S, Vn.2S, Vm.2S |