4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or https://opensource.org/licenses/CDDL-1.0.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
22 * Copyright (C) 2016 Gvozden Nešković. All rights reserved.
25 #if defined(__x86_64) && defined(HAVE_AVX512F)
27 #include <sys/byteorder.h>
28 #include <sys/frame.h>
29 #include <sys/spa_checksum.h>
30 #include <sys/string.h>
32 #include <zfs_fletcher.h>
35 #define __asm __asm__ __volatile__
39 fletcher_4_avx512f_init(fletcher_4_ctx_t
*ctx
)
41 memset(ctx
->avx512
, 0, 4 * sizeof (zfs_fletcher_avx512_t
));
45 fletcher_4_avx512f_fini(fletcher_4_ctx_t
*ctx
, zio_cksum_t
*zcp
)
48 CcA
[] = { 0, 0, 1, 3, 6, 10, 15, 21 },
49 CcB
[] = { 28, 36, 44, 52, 60, 68, 76, 84 },
50 DcA
[] = { 0, 0, 0, 1, 4, 10, 20, 35 },
51 DcB
[] = { 56, 84, 120, 164, 216, 276, 344, 420 },
52 DcC
[] = { 448, 512, 576, 640, 704, 768, 832, 896 };
57 A
= ctx
->avx512
[0].v
[0];
58 B
= 8 * ctx
->avx512
[1].v
[0];
59 C
= 64 * ctx
->avx512
[2].v
[0] - CcB
[0] * ctx
->avx512
[1].v
[0];
60 D
= 512 * ctx
->avx512
[3].v
[0] - DcC
[0] * ctx
->avx512
[2].v
[0] +
61 DcB
[0] * ctx
->avx512
[1].v
[0];
63 for (i
= 1; i
< 8; i
++) {
64 A
+= ctx
->avx512
[0].v
[i
];
65 B
+= 8 * ctx
->avx512
[1].v
[i
] - i
* ctx
->avx512
[0].v
[i
];
66 C
+= 64 * ctx
->avx512
[2].v
[i
] - CcB
[i
] * ctx
->avx512
[1].v
[i
] +
67 CcA
[i
] * ctx
->avx512
[0].v
[i
];
68 D
+= 512 * ctx
->avx512
[3].v
[i
] - DcC
[i
] * ctx
->avx512
[2].v
[i
] +
69 DcB
[i
] * ctx
->avx512
[1].v
[i
] - DcA
[i
] * ctx
->avx512
[0].v
[i
];
72 ZIO_SET_CHECKSUM(zcp
, A
, B
, C
, D
);
75 #define FLETCHER_4_AVX512_RESTORE_CTX(ctx) \
77 __asm("vmovdqu64 %0, %%zmm0" :: "m" ((ctx)->avx512[0])); \
78 __asm("vmovdqu64 %0, %%zmm1" :: "m" ((ctx)->avx512[1])); \
79 __asm("vmovdqu64 %0, %%zmm2" :: "m" ((ctx)->avx512[2])); \
80 __asm("vmovdqu64 %0, %%zmm3" :: "m" ((ctx)->avx512[3])); \
83 #define FLETCHER_4_AVX512_SAVE_CTX(ctx) \
85 __asm("vmovdqu64 %%zmm0, %0" : "=m" ((ctx)->avx512[0])); \
86 __asm("vmovdqu64 %%zmm1, %0" : "=m" ((ctx)->avx512[1])); \
87 __asm("vmovdqu64 %%zmm2, %0" : "=m" ((ctx)->avx512[2])); \
88 __asm("vmovdqu64 %%zmm3, %0" : "=m" ((ctx)->avx512[3])); \
92 fletcher_4_avx512f_native(fletcher_4_ctx_t
*ctx
, const void *buf
, uint64_t size
)
94 const uint32_t *ip
= buf
;
95 const uint32_t *ipend
= (uint32_t *)((uint8_t *)ip
+ size
);
97 FLETCHER_4_AVX512_RESTORE_CTX(ctx
);
100 __asm("vpmovzxdq %0, %%zmm4"::"m" (*ip
));
101 __asm("vpaddq %zmm4, %zmm0, %zmm0");
102 __asm("vpaddq %zmm0, %zmm1, %zmm1");
103 __asm("vpaddq %zmm1, %zmm2, %zmm2");
104 __asm("vpaddq %zmm2, %zmm3, %zmm3");
105 } while ((ip
+= 8) < ipend
);
107 FLETCHER_4_AVX512_SAVE_CTX(ctx
);
109 STACK_FRAME_NON_STANDARD(fletcher_4_avx512f_native
);
112 fletcher_4_avx512f_byteswap(fletcher_4_ctx_t
*ctx
, const void *buf
,
115 static const uint64_t byteswap_mask
= 0xFFULL
;
116 const uint32_t *ip
= buf
;
117 const uint32_t *ipend
= (uint32_t *)((uint8_t *)ip
+ size
);
119 FLETCHER_4_AVX512_RESTORE_CTX(ctx
);
121 __asm("vpbroadcastq %0, %%zmm8" :: "r" (byteswap_mask
));
122 __asm("vpsllq $8, %zmm8, %zmm9");
123 __asm("vpsllq $16, %zmm8, %zmm10");
124 __asm("vpsllq $24, %zmm8, %zmm11");
127 __asm("vpmovzxdq %0, %%zmm5"::"m" (*ip
));
129 __asm("vpsrlq $24, %zmm5, %zmm6");
130 __asm("vpandd %zmm8, %zmm6, %zmm6");
131 __asm("vpsrlq $8, %zmm5, %zmm7");
132 __asm("vpandd %zmm9, %zmm7, %zmm7");
133 __asm("vpord %zmm6, %zmm7, %zmm4");
134 __asm("vpsllq $8, %zmm5, %zmm6");
135 __asm("vpandd %zmm10, %zmm6, %zmm6");
136 __asm("vpord %zmm6, %zmm4, %zmm4");
137 __asm("vpsllq $24, %zmm5, %zmm5");
138 __asm("vpandd %zmm11, %zmm5, %zmm5");
139 __asm("vpord %zmm5, %zmm4, %zmm4");
141 __asm("vpaddq %zmm4, %zmm0, %zmm0");
142 __asm("vpaddq %zmm0, %zmm1, %zmm1");
143 __asm("vpaddq %zmm1, %zmm2, %zmm2");
144 __asm("vpaddq %zmm2, %zmm3, %zmm3");
145 } while ((ip
+= 8) < ipend
);
147 FLETCHER_4_AVX512_SAVE_CTX(ctx
)
149 STACK_FRAME_NON_STANDARD(fletcher_4_avx512f_byteswap
);
152 fletcher_4_avx512f_valid(void)
154 return (kfpu_allowed() && zfs_avx512f_available());
157 const fletcher_4_ops_t fletcher_4_avx512f_ops
= {
158 .init_native
= fletcher_4_avx512f_init
,
159 .fini_native
= fletcher_4_avx512f_fini
,
160 .compute_native
= fletcher_4_avx512f_native
,
161 .init_byteswap
= fletcher_4_avx512f_init
,
162 .fini_byteswap
= fletcher_4_avx512f_fini
,
163 .compute_byteswap
= fletcher_4_avx512f_byteswap
,
164 .valid
= fletcher_4_avx512f_valid
,
169 #if defined(HAVE_AVX512BW)
171 fletcher_4_avx512bw_byteswap(fletcher_4_ctx_t
*ctx
, const void *buf
,
174 static const zfs_fletcher_avx512_t mask
= {
175 .v
= { 0xFFFFFFFF00010203, 0xFFFFFFFF08090A0B,
176 0xFFFFFFFF00010203, 0xFFFFFFFF08090A0B,
177 0xFFFFFFFF00010203, 0xFFFFFFFF08090A0B,
178 0xFFFFFFFF00010203, 0xFFFFFFFF08090A0B }
180 const uint32_t *ip
= buf
;
181 const uint32_t *ipend
= (uint32_t *)((uint8_t *)ip
+ size
);
183 FLETCHER_4_AVX512_RESTORE_CTX(ctx
);
185 __asm("vmovdqu64 %0, %%zmm5" :: "m" (mask
));
188 __asm("vpmovzxdq %0, %%zmm4"::"m" (*ip
));
190 __asm("vpshufb %zmm5, %zmm4, %zmm4");
192 __asm("vpaddq %zmm4, %zmm0, %zmm0");
193 __asm("vpaddq %zmm0, %zmm1, %zmm1");
194 __asm("vpaddq %zmm1, %zmm2, %zmm2");
195 __asm("vpaddq %zmm2, %zmm3, %zmm3");
196 } while ((ip
+= 8) < ipend
);
198 FLETCHER_4_AVX512_SAVE_CTX(ctx
)
200 STACK_FRAME_NON_STANDARD(fletcher_4_avx512bw_byteswap
);
203 fletcher_4_avx512bw_valid(void)
205 return (fletcher_4_avx512f_valid() && zfs_avx512bw_available());
208 const fletcher_4_ops_t fletcher_4_avx512bw_ops
= {
209 .init_native
= fletcher_4_avx512f_init
,
210 .fini_native
= fletcher_4_avx512f_fini
,
211 .compute_native
= fletcher_4_avx512f_native
,
212 .init_byteswap
= fletcher_4_avx512f_init
,
213 .fini_byteswap
= fletcher_4_avx512f_fini
,
214 .compute_byteswap
= fletcher_4_avx512bw_byteswap
,
215 .valid
= fletcher_4_avx512bw_valid
,
221 #endif /* defined(__x86_64) && defined(HAVE_AVX512F) */