|  | #ifndef _ASM_X86_XOR_64_H | 
|  | #define _ASM_X86_XOR_64_H | 
|  |  | 
|  | static struct xor_block_template xor_block_sse = { | 
|  | .name = "generic_sse", | 
|  | .do_2 = xor_sse_2, | 
|  | .do_3 = xor_sse_3, | 
|  | .do_4 = xor_sse_4, | 
|  | .do_5 = xor_sse_5, | 
|  | }; | 
|  |  | 
|  |  | 
|  | /* Also try the AVX routines */ | 
|  | #include <asm/xor_avx.h> | 
|  |  | 
|  | /* We force the use of the SSE xor block because it can write around L2. | 
|  | We may also be able to load into the L1 only depending on how the cpu | 
|  | deals with a load to a line that is being prefetched.  */ | 
|  | #undef XOR_TRY_TEMPLATES | 
|  | #define XOR_TRY_TEMPLATES			\ | 
|  | do {						\ | 
|  | AVX_XOR_SPEED;				\ | 
|  | xor_speed(&xor_block_sse_pf64);		\ | 
|  | xor_speed(&xor_block_sse);		\ | 
|  | } while (0) | 
|  |  | 
|  | #endif /* _ASM_X86_XOR_64_H */ |