@@ -169,7 +169,7 @@ pub unsafe fn _mm_maskz_expandloadu_epi8(k: __mmask16, mem_addr: *const i8) -> _
169169#[ target_feature( enable = "avx512vbmi2" ) ]
170170#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
171171#[ cfg_attr( test, assert_instr( vpcompressw) ) ]
172- pub unsafe fn _mm512_mask_compressstoreu_epi16 ( base_addr : * mut u8 , k : __mmask32 , a : __m512i ) {
172+ pub unsafe fn _mm512_mask_compressstoreu_epi16 ( base_addr : * mut i16 , k : __mmask32 , a : __m512i ) {
173173 vcompressstorew ( base_addr as * mut _ , a. as_i16x32 ( ) , k)
174174}
175175
@@ -180,7 +180,7 @@ pub unsafe fn _mm512_mask_compressstoreu_epi16(base_addr: *mut u8, k: __mmask32,
180180#[ target_feature( enable = "avx512vbmi2,avx512vl" ) ]
181181#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
182182#[ cfg_attr( test, assert_instr( vpcompressw) ) ]
183- pub unsafe fn _mm256_mask_compressstoreu_epi16 ( base_addr : * mut u8 , k : __mmask16 , a : __m256i ) {
183+ pub unsafe fn _mm256_mask_compressstoreu_epi16 ( base_addr : * mut i16 , k : __mmask16 , a : __m256i ) {
184184 vcompressstorew256 ( base_addr as * mut _ , a. as_i16x16 ( ) , k)
185185}
186186
@@ -191,7 +191,7 @@ pub unsafe fn _mm256_mask_compressstoreu_epi16(base_addr: *mut u8, k: __mmask16,
191191#[ target_feature( enable = "avx512vbmi2,avx512vl" ) ]
192192#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
193193#[ cfg_attr( test, assert_instr( vpcompressw) ) ]
194- pub unsafe fn _mm_mask_compressstoreu_epi16 ( base_addr : * mut u8 , k : __mmask8 , a : __m128i ) {
194+ pub unsafe fn _mm_mask_compressstoreu_epi16 ( base_addr : * mut i16 , k : __mmask8 , a : __m128i ) {
195195 vcompressstorew128 ( base_addr as * mut _ , a. as_i16x8 ( ) , k)
196196}
197197
@@ -202,8 +202,8 @@ pub unsafe fn _mm_mask_compressstoreu_epi16(base_addr: *mut u8, k: __mmask8, a:
202202#[ target_feature( enable = "avx512vbmi2" ) ]
203203#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
204204#[ cfg_attr( test, assert_instr( vpcompressb) ) ]
205- pub unsafe fn _mm512_mask_compressstoreu_epi8 ( base_addr : * mut u8 , k : __mmask64 , a : __m512i ) {
206- vcompressstoreb ( base_addr as * mut _ , a. as_i8x64 ( ) , k)
205+ pub unsafe fn _mm512_mask_compressstoreu_epi8 ( base_addr : * mut i8 , k : __mmask64 , a : __m512i ) {
206+ vcompressstoreb ( base_addr, a. as_i8x64 ( ) , k)
207207}
208208
209209/// Contiguously store the active 8-bit integers in a (those with their respective bit set in writemask k) to unaligned memory at base_addr.
@@ -213,8 +213,8 @@ pub unsafe fn _mm512_mask_compressstoreu_epi8(base_addr: *mut u8, k: __mmask64,
213213#[ target_feature( enable = "avx512vbmi2,avx512vl" ) ]
214214#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
215215#[ cfg_attr( test, assert_instr( vpcompressb) ) ]
216- pub unsafe fn _mm256_mask_compressstoreu_epi8 ( base_addr : * mut u8 , k : __mmask32 , a : __m256i ) {
217- vcompressstoreb256 ( base_addr as * mut _ , a. as_i8x32 ( ) , k)
216+ pub unsafe fn _mm256_mask_compressstoreu_epi8 ( base_addr : * mut i8 , k : __mmask32 , a : __m256i ) {
217+ vcompressstoreb256 ( base_addr, a. as_i8x32 ( ) , k)
218218}
219219
220220/// Contiguously store the active 8-bit integers in a (those with their respective bit set in writemask k) to unaligned memory at base_addr.
@@ -224,8 +224,8 @@ pub unsafe fn _mm256_mask_compressstoreu_epi8(base_addr: *mut u8, k: __mmask32,
224224#[ target_feature( enable = "avx512vbmi2,avx512vl" ) ]
225225#[ unstable( feature = "stdarch_x86_avx512" , issue = "111137" ) ]
226226#[ cfg_attr( test, assert_instr( vpcompressb) ) ]
227- pub unsafe fn _mm_mask_compressstoreu_epi8 ( base_addr : * mut u8 , k : __mmask16 , a : __m128i ) {
228- vcompressstoreb128 ( base_addr as * mut _ , a. as_i8x16 ( ) , k)
227+ pub unsafe fn _mm_mask_compressstoreu_epi8 ( base_addr : * mut i8 , k : __mmask16 , a : __m128i ) {
228+ vcompressstoreb128 ( base_addr, a. as_i8x16 ( ) , k)
229229}
230230
231231/// Contiguously store the active 16-bit integers in a (those with their respective bit set in writemask k) to dst, and pass through the remaining elements from src.
@@ -3853,13 +3853,9 @@ mod tests {
38533853 10 , 9 , 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ,
38543854 ) ;
38553855 let mut r = [ 0_i16 ; 32 ] ;
3856- _mm512_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3856+ _mm512_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0 , a) ;
38573857 assert_eq ! ( & r, & [ 0_i16 ; 32 ] ) ;
3858- _mm512_mask_compressstoreu_epi16 (
3859- r. as_mut_ptr ( ) as * mut _ ,
3860- 0b11110000_11001010_11111111_00000000 ,
3861- a,
3862- ) ;
3858+ _mm512_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0b11110000_11001010_11111111_00000000 , a) ;
38633859 assert_eq ! (
38643860 & r,
38653861 & [
@@ -3873,19 +3869,19 @@ mod tests {
38733869 unsafe fn test_mm256_mask_compressstoreu_epi16 ( ) {
38743870 let a = _mm256_set_epi16 ( 16 , 15 , 14 , 13 , 12 , 11 , 10 , 9 , 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ) ;
38753871 let mut r = [ 0_i16 ; 16 ] ;
3876- _mm256_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3872+ _mm256_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0 , a) ;
38773873 assert_eq ! ( & r, & [ 0_i16 ; 16 ] ) ;
3878- _mm256_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) as * mut _ , 0b11110000_11001010 , a) ;
3874+ _mm256_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0b11110000_11001010 , a) ;
38793875 assert_eq ! ( & r, & [ 2 , 4 , 7 , 8 , 13 , 14 , 15 , 16 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ] ) ;
38803876 }
38813877
38823878 #[ simd_test( enable = "avx512vbmi2,avx512vl" ) ]
38833879 unsafe fn test_mm_mask_compressstoreu_epi16 ( ) {
38843880 let a = _mm_set_epi16 ( 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ) ;
38853881 let mut r = [ 0_i16 ; 8 ] ;
3886- _mm_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3882+ _mm_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0 , a) ;
38873883 assert_eq ! ( & r, & [ 0_i16 ; 8 ] ) ;
3888- _mm_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) as * mut _ , 0b11110000 , a) ;
3884+ _mm_mask_compressstoreu_epi16 ( r. as_mut_ptr ( ) , 0b11110000 , a) ;
38893885 assert_eq ! ( & r, & [ 5 , 6 , 7 , 8 , 0 , 0 , 0 , 0 ] ) ;
38903886 }
38913887
@@ -3897,10 +3893,10 @@ mod tests {
38973893 20 , 19 , 18 , 17 , 16 , 15 , 14 , 13 , 12 , 11 , 10 , 9 , 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ,
38983894 ) ;
38993895 let mut r = [ 0_i8 ; 64 ] ;
3900- _mm512_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3896+ _mm512_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) , 0 , a) ;
39013897 assert_eq ! ( & r, & [ 0_i8 ; 64 ] ) ;
39023898 _mm512_mask_compressstoreu_epi8 (
3903- r. as_mut_ptr ( ) as * mut _ ,
3899+ r. as_mut_ptr ( ) ,
39043900 0b11110000_11001010_11111111_00000000_10101010_01010101_11110000_00001111 ,
39053901 a,
39063902 ) ;
@@ -3921,13 +3917,9 @@ mod tests {
39213917 10 , 9 , 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ,
39223918 ) ;
39233919 let mut r = [ 0_i8 ; 32 ] ;
3924- _mm256_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3920+ _mm256_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) , 0 , a) ;
39253921 assert_eq ! ( & r, & [ 0_i8 ; 32 ] ) ;
3926- _mm256_mask_compressstoreu_epi8 (
3927- r. as_mut_ptr ( ) as * mut _ ,
3928- 0b11110000_11001010_11111111_00000000 ,
3929- a,
3930- ) ;
3922+ _mm256_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) , 0b11110000_11001010_11111111_00000000 , a) ;
39313923 assert_eq ! (
39323924 & r,
39333925 & [
@@ -3941,9 +3933,9 @@ mod tests {
39413933 unsafe fn test_mm_mask_compressstoreu_epi8 ( ) {
39423934 let a = _mm_set_epi8 ( 16 , 15 , 14 , 13 , 12 , 11 , 10 , 9 , 8 , 7 , 6 , 5 , 4 , 3 , 2 , 1 ) ;
39433935 let mut r = [ 0_i8 ; 16 ] ;
3944- _mm_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) as * mut _ , 0 , a) ;
3936+ _mm_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) , 0 , a) ;
39453937 assert_eq ! ( & r, & [ 0_i8 ; 16 ] ) ;
3946- _mm_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) as * mut _ , 0b11110000_11001010 , a) ;
3938+ _mm_mask_compressstoreu_epi8 ( r. as_mut_ptr ( ) , 0b11110000_11001010 , a) ;
39473939 assert_eq ! ( & r, & [ 2 , 4 , 7 , 8 , 13 , 14 , 15 , 16 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ] ) ;
39483940 }
39493941}
0 commit comments