First of all, it had a failure in SEED: LTC_KSEED failed for x=0, I got: expected actual (ciphertext) 5e == 5e ba == ba c6 == c6 e0 == e0 05 != 00 4e != 00 16 != 00 68 != 00 19 == 19 af == af f1 == f1 cc == cc 6d != 00 34 != 00 6c != 00 db != 00 Since SEED uses the 32H macros, this is really analogous to the problem I saw with the 64H macros in Camellia with gcc. Not sure why gcc only had a problem with 64H and not 32H, but since this is an interaction with the optimizer, it's not going to happen every time the macro is used (hence why the store tests pass; only when you get into the complexity of a real cipher do you start having problems) and it makes sense it will vary from compiler to compiler. Anyway, I went ahead and added the ability to use __builtin_bswap32, in addition to __builtin_bswap64, which I already did in a previous commit. This solves the problem for clang, although I had to add new logic to detect the bswap builtins in clang, since it has a different way to detect them than gcc (see the comments in the code). The detection logic was complicated enough, and applied to both the 32H and 64H macros, so I factored out the detection logic into tomcrypt_cfg.h.
444 lines
15 KiB
C
444 lines
15 KiB
C
/* fix for MSVC ...evil! */
|
|
#ifdef _MSC_VER
|
|
#define CONST64(n) n ## ui64
|
|
typedef unsigned __int64 ulong64;
|
|
#else
|
|
#define CONST64(n) n ## ULL
|
|
typedef unsigned long long ulong64;
|
|
#endif
|
|
|
|
/* this is the "32-bit at least" data type
|
|
* Re-define it to suit your platform but it must be at least 32-bits
|
|
*/
|
|
#if defined(__x86_64__) || (defined(__sparc__) && defined(__arch64__))
|
|
typedef unsigned ulong32;
|
|
#else
|
|
typedef unsigned long ulong32;
|
|
#endif
|
|
|
|
/* ---- HELPER MACROS ---- */
|
|
#ifdef ENDIAN_NEUTRAL
|
|
|
|
#define STORE32L(x, y) \
|
|
{ (y)[3] = (unsigned char)(((x)>>24)&255); (y)[2] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[1] = (unsigned char)(((x)>>8)&255); (y)[0] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD32L(x, y) \
|
|
{ x = ((unsigned long)((y)[3] & 255)<<24) | \
|
|
((unsigned long)((y)[2] & 255)<<16) | \
|
|
((unsigned long)((y)[1] & 255)<<8) | \
|
|
((unsigned long)((y)[0] & 255)); }
|
|
|
|
#define STORE64L(x, y) \
|
|
{ (y)[7] = (unsigned char)(((x)>>56)&255); (y)[6] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[5] = (unsigned char)(((x)>>40)&255); (y)[4] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[3] = (unsigned char)(((x)>>24)&255); (y)[2] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[1] = (unsigned char)(((x)>>8)&255); (y)[0] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64L(x, y) \
|
|
{ x = (((ulong64)((y)[7] & 255))<<56)|(((ulong64)((y)[6] & 255))<<48)| \
|
|
(((ulong64)((y)[5] & 255))<<40)|(((ulong64)((y)[4] & 255))<<32)| \
|
|
(((ulong64)((y)[3] & 255))<<24)|(((ulong64)((y)[2] & 255))<<16)| \
|
|
(((ulong64)((y)[1] & 255))<<8)|(((ulong64)((y)[0] & 255))); }
|
|
|
|
#define STORE32H(x, y) \
|
|
{ (y)[0] = (unsigned char)(((x)>>24)&255); (y)[1] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[2] = (unsigned char)(((x)>>8)&255); (y)[3] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD32H(x, y) \
|
|
{ x = ((unsigned long)((y)[0] & 255)<<24) | \
|
|
((unsigned long)((y)[1] & 255)<<16) | \
|
|
((unsigned long)((y)[2] & 255)<<8) | \
|
|
((unsigned long)((y)[3] & 255)); }
|
|
|
|
#define STORE64H(x, y) \
|
|
{ (y)[0] = (unsigned char)(((x)>>56)&255); (y)[1] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[2] = (unsigned char)(((x)>>40)&255); (y)[3] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[4] = (unsigned char)(((x)>>24)&255); (y)[5] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[6] = (unsigned char)(((x)>>8)&255); (y)[7] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64H(x, y) \
|
|
{ x = (((ulong64)((y)[0] & 255))<<56)|(((ulong64)((y)[1] & 255))<<48) | \
|
|
(((ulong64)((y)[2] & 255))<<40)|(((ulong64)((y)[3] & 255))<<32) | \
|
|
(((ulong64)((y)[4] & 255))<<24)|(((ulong64)((y)[5] & 255))<<16) | \
|
|
(((ulong64)((y)[6] & 255))<<8)|(((ulong64)((y)[7] & 255))); }
|
|
|
|
#endif /* ENDIAN_NEUTRAL */
|
|
|
|
#ifdef ENDIAN_LITTLE
|
|
|
|
#ifdef LTC_HAVE_BSWAP_BUILTIN
|
|
|
|
#define STORE32H(x, y) \
|
|
{ ulong32 __t = __builtin_bswap32 ((x)); \
|
|
XMEMCPY ((y), &__t, 4); }
|
|
|
|
#define LOAD32H(x, y) \
|
|
{ XMEMCPY (&(x), (y), 4); \
|
|
(x) = __builtin_bswap32 ((x)); }
|
|
|
|
#elif !defined(LTC_NO_BSWAP) && (defined(INTEL_CC) || (defined(__GNUC__) && (defined(__DJGPP__) || defined(__CYGWIN__) || defined(__MINGW32__) || defined(__i386__) || defined(__x86_64__))))
|
|
|
|
#define STORE32H(x, y) \
|
|
asm __volatile__ ( \
|
|
"bswapl %0 \n\t" \
|
|
"movl %0,(%1)\n\t" \
|
|
"bswapl %0 \n\t" \
|
|
::"r"(x), "r"(y));
|
|
|
|
#define LOAD32H(x, y) \
|
|
asm __volatile__ ( \
|
|
"movl (%1),%0\n\t" \
|
|
"bswapl %0\n\t" \
|
|
:"=r"(x): "r"(y));
|
|
|
|
#else
|
|
|
|
#define STORE32H(x, y) \
|
|
{ (y)[0] = (unsigned char)(((x)>>24)&255); (y)[1] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[2] = (unsigned char)(((x)>>8)&255); (y)[3] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD32H(x, y) \
|
|
{ x = ((unsigned long)((y)[0] & 255)<<24) | \
|
|
((unsigned long)((y)[1] & 255)<<16) | \
|
|
((unsigned long)((y)[2] & 255)<<8) | \
|
|
((unsigned long)((y)[3] & 255)); }
|
|
|
|
#endif
|
|
|
|
#ifdef LTC_HAVE_BSWAP_BUILTIN
|
|
|
|
#define STORE64H(x, y) \
|
|
{ ulong64 __t = __builtin_bswap64 ((x)); \
|
|
XMEMCPY ((y), &__t, 8); }
|
|
|
|
#define LOAD64H(x, y) \
|
|
{ XMEMCPY (&(x), (y), 8); \
|
|
(x) = __builtin_bswap64 ((x)); }
|
|
|
|
/* x86_64 processor */
|
|
#elif !defined(LTC_NO_BSWAP) && (defined(__GNUC__) && defined(__x86_64__))
|
|
|
|
#define STORE64H(x, y) \
|
|
asm __volatile__ ( \
|
|
"bswapq %0 \n\t" \
|
|
"movq %0,(%1)\n\t" \
|
|
"bswapq %0 \n\t" \
|
|
::"r"(x), "r"(y): "memory");
|
|
|
|
#define LOAD64H(x, y) \
|
|
asm __volatile__ ( \
|
|
"movq (%1),%0\n\t" \
|
|
"bswapq %0\n\t" \
|
|
:"=r"(x): "r"(y): "memory");
|
|
|
|
#else
|
|
|
|
#define STORE64H(x, y) \
|
|
{ (y)[0] = (unsigned char)(((x)>>56)&255); (y)[1] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[2] = (unsigned char)(((x)>>40)&255); (y)[3] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[4] = (unsigned char)(((x)>>24)&255); (y)[5] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[6] = (unsigned char)(((x)>>8)&255); (y)[7] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64H(x, y) \
|
|
{ x = (((ulong64)((y)[0] & 255))<<56)|(((ulong64)((y)[1] & 255))<<48) | \
|
|
(((ulong64)((y)[2] & 255))<<40)|(((ulong64)((y)[3] & 255))<<32) | \
|
|
(((ulong64)((y)[4] & 255))<<24)|(((ulong64)((y)[5] & 255))<<16) | \
|
|
(((ulong64)((y)[6] & 255))<<8)|(((ulong64)((y)[7] & 255))); }
|
|
|
|
#endif
|
|
|
|
#ifdef ENDIAN_32BITWORD
|
|
|
|
#define STORE32L(x, y) \
|
|
{ ulong32 __t = (x); XMEMCPY(y, &__t, 4); }
|
|
|
|
#define LOAD32L(x, y) \
|
|
XMEMCPY(&(x), y, 4);
|
|
|
|
#define STORE64L(x, y) \
|
|
{ (y)[7] = (unsigned char)(((x)>>56)&255); (y)[6] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[5] = (unsigned char)(((x)>>40)&255); (y)[4] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[3] = (unsigned char)(((x)>>24)&255); (y)[2] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[1] = (unsigned char)(((x)>>8)&255); (y)[0] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64L(x, y) \
|
|
{ x = (((ulong64)((y)[7] & 255))<<56)|(((ulong64)((y)[6] & 255))<<48)| \
|
|
(((ulong64)((y)[5] & 255))<<40)|(((ulong64)((y)[4] & 255))<<32)| \
|
|
(((ulong64)((y)[3] & 255))<<24)|(((ulong64)((y)[2] & 255))<<16)| \
|
|
(((ulong64)((y)[1] & 255))<<8)|(((ulong64)((y)[0] & 255))); }
|
|
|
|
#else /* 64-bit words then */
|
|
|
|
#define STORE32L(x, y) \
|
|
{ ulong32 __t = (x); XMEMCPY(y, &__t, 4); }
|
|
|
|
#define LOAD32L(x, y) \
|
|
{ XMEMCPY(&(x), y, 4); x &= 0xFFFFFFFF; }
|
|
|
|
#define STORE64L(x, y) \
|
|
{ ulong64 __t = (x); XMEMCPY(y, &__t, 8); }
|
|
|
|
#define LOAD64L(x, y) \
|
|
{ XMEMCPY(&(x), y, 8); }
|
|
|
|
#endif /* ENDIAN_64BITWORD */
|
|
|
|
#endif /* ENDIAN_LITTLE */
|
|
|
|
#ifdef ENDIAN_BIG
|
|
#define STORE32L(x, y) \
|
|
{ (y)[3] = (unsigned char)(((x)>>24)&255); (y)[2] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[1] = (unsigned char)(((x)>>8)&255); (y)[0] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD32L(x, y) \
|
|
{ x = ((unsigned long)((y)[3] & 255)<<24) | \
|
|
((unsigned long)((y)[2] & 255)<<16) | \
|
|
((unsigned long)((y)[1] & 255)<<8) | \
|
|
((unsigned long)((y)[0] & 255)); }
|
|
|
|
#define STORE64L(x, y) \
|
|
{ (y)[7] = (unsigned char)(((x)>>56)&255); (y)[6] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[5] = (unsigned char)(((x)>>40)&255); (y)[4] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[3] = (unsigned char)(((x)>>24)&255); (y)[2] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[1] = (unsigned char)(((x)>>8)&255); (y)[0] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64L(x, y) \
|
|
{ x = (((ulong64)((y)[7] & 255))<<56)|(((ulong64)((y)[6] & 255))<<48) | \
|
|
(((ulong64)((y)[5] & 255))<<40)|(((ulong64)((y)[4] & 255))<<32) | \
|
|
(((ulong64)((y)[3] & 255))<<24)|(((ulong64)((y)[2] & 255))<<16) | \
|
|
(((ulong64)((y)[1] & 255))<<8)|(((ulong64)((y)[0] & 255))); }
|
|
|
|
#ifdef ENDIAN_32BITWORD
|
|
|
|
#define STORE32H(x, y) \
|
|
{ ulong32 __t = (x); XMEMCPY(y, &__t, 4); }
|
|
|
|
#define LOAD32H(x, y) \
|
|
XMEMCPY(&(x), y, 4);
|
|
|
|
#define STORE64H(x, y) \
|
|
{ (y)[0] = (unsigned char)(((x)>>56)&255); (y)[1] = (unsigned char)(((x)>>48)&255); \
|
|
(y)[2] = (unsigned char)(((x)>>40)&255); (y)[3] = (unsigned char)(((x)>>32)&255); \
|
|
(y)[4] = (unsigned char)(((x)>>24)&255); (y)[5] = (unsigned char)(((x)>>16)&255); \
|
|
(y)[6] = (unsigned char)(((x)>>8)&255); (y)[7] = (unsigned char)((x)&255); }
|
|
|
|
#define LOAD64H(x, y) \
|
|
{ x = (((ulong64)((y)[0] & 255))<<56)|(((ulong64)((y)[1] & 255))<<48)| \
|
|
(((ulong64)((y)[2] & 255))<<40)|(((ulong64)((y)[3] & 255))<<32)| \
|
|
(((ulong64)((y)[4] & 255))<<24)|(((ulong64)((y)[5] & 255))<<16)| \
|
|
(((ulong64)((y)[6] & 255))<<8)| (((ulong64)((y)[7] & 255))); }
|
|
|
|
#else /* 64-bit words then */
|
|
|
|
#define STORE32H(x, y) \
|
|
{ ulong32 __t = (x); XMEMCPY(y, &__t, 4); }
|
|
|
|
#define LOAD32H(x, y) \
|
|
{ XMEMCPY(&(x), y, 4); x &= 0xFFFFFFFF; }
|
|
|
|
#define STORE64H(x, y) \
|
|
{ ulong64 __t = (x); XMEMCPY(y, &__t, 8); }
|
|
|
|
#define LOAD64H(x, y) \
|
|
{ XMEMCPY(&(x), y, 8); }
|
|
|
|
#endif /* ENDIAN_64BITWORD */
|
|
#endif /* ENDIAN_BIG */
|
|
|
|
#define BSWAP(x) ( ((x>>24)&0x000000FFUL) | ((x<<24)&0xFF000000UL) | \
|
|
((x>>8)&0x0000FF00UL) | ((x<<8)&0x00FF0000UL) )
|
|
|
|
|
|
/* 32-bit Rotates */
|
|
#if defined(_MSC_VER)
|
|
|
|
/* instrinsic rotate */
|
|
#include <stdlib.h>
|
|
#pragma intrinsic(_lrotr,_lrotl)
|
|
#define ROR(x,n) _lrotr(x,n)
|
|
#define ROL(x,n) _lrotl(x,n)
|
|
#define RORc(x,n) _lrotr(x,n)
|
|
#define ROLc(x,n) _lrotl(x,n)
|
|
|
|
#elif !defined(__STRICT_ANSI__) && defined(__GNUC__) && (defined(__i386__) || defined(__x86_64__)) && !defined(INTEL_CC) && !defined(LTC_NO_ASM)
|
|
|
|
static inline unsigned ROL(unsigned word, int i)
|
|
{
|
|
asm ("roll %%cl,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"c" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned ROR(unsigned word, int i)
|
|
{
|
|
asm ("rorl %%cl,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"c" (i));
|
|
return word;
|
|
}
|
|
|
|
#ifndef LTC_NO_ROLC
|
|
|
|
static inline unsigned ROLc(unsigned word, const int i)
|
|
{
|
|
asm ("roll %2,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"I" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned RORc(unsigned word, const int i)
|
|
{
|
|
asm ("rorl %2,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"I" (i));
|
|
return word;
|
|
}
|
|
|
|
#else
|
|
|
|
#define ROLc ROL
|
|
#define RORc ROR
|
|
|
|
#endif
|
|
|
|
#elif !defined(__STRICT_ANSI__) && defined(LTC_PPC32)
|
|
|
|
static inline unsigned ROL(unsigned word, int i)
|
|
{
|
|
asm ("rotlw %0,%0,%2"
|
|
:"=r" (word)
|
|
:"0" (word),"r" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned ROR(unsigned word, int i)
|
|
{
|
|
asm ("rotlw %0,%0,%2"
|
|
:"=r" (word)
|
|
:"0" (word),"r" (32-i));
|
|
return word;
|
|
}
|
|
|
|
#ifndef LTC_NO_ROLC
|
|
|
|
static inline unsigned ROLc(unsigned word, const int i)
|
|
{
|
|
asm ("rotlwi %0,%0,%2"
|
|
:"=r" (word)
|
|
:"0" (word),"I" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned RORc(unsigned word, const int i)
|
|
{
|
|
asm ("rotrwi %0,%0,%2"
|
|
:"=r" (word)
|
|
:"0" (word),"I" (i));
|
|
return word;
|
|
}
|
|
|
|
#else
|
|
|
|
#define ROLc ROL
|
|
#define RORc ROR
|
|
|
|
#endif
|
|
|
|
|
|
#else
|
|
|
|
/* rotates the hard way */
|
|
#define ROL(x, y) ( (((unsigned long)(x)<<(unsigned long)((y)&31)) | (((unsigned long)(x)&0xFFFFFFFFUL)>>(unsigned long)(32-((y)&31)))) & 0xFFFFFFFFUL)
|
|
#define ROR(x, y) ( ((((unsigned long)(x)&0xFFFFFFFFUL)>>(unsigned long)((y)&31)) | ((unsigned long)(x)<<(unsigned long)(32-((y)&31)))) & 0xFFFFFFFFUL)
|
|
#define ROLc(x, y) ( (((unsigned long)(x)<<(unsigned long)((y)&31)) | (((unsigned long)(x)&0xFFFFFFFFUL)>>(unsigned long)(32-((y)&31)))) & 0xFFFFFFFFUL)
|
|
#define RORc(x, y) ( ((((unsigned long)(x)&0xFFFFFFFFUL)>>(unsigned long)((y)&31)) | ((unsigned long)(x)<<(unsigned long)(32-((y)&31)))) & 0xFFFFFFFFUL)
|
|
|
|
#endif
|
|
|
|
|
|
/* 64-bit Rotates */
|
|
#if !defined(__STRICT_ANSI__) && defined(__GNUC__) && defined(__x86_64__) && !defined(LTC_NO_ASM)
|
|
|
|
static inline unsigned long ROL64(unsigned long word, int i)
|
|
{
|
|
asm("rolq %%cl,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"c" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned long ROR64(unsigned long word, int i)
|
|
{
|
|
asm("rorq %%cl,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"c" (i));
|
|
return word;
|
|
}
|
|
|
|
#ifndef LTC_NO_ROLC
|
|
|
|
static inline unsigned long ROL64c(unsigned long word, const int i)
|
|
{
|
|
asm("rolq %2,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"J" (i));
|
|
return word;
|
|
}
|
|
|
|
static inline unsigned long ROR64c(unsigned long word, const int i)
|
|
{
|
|
asm("rorq %2,%0"
|
|
:"=r" (word)
|
|
:"0" (word),"J" (i));
|
|
return word;
|
|
}
|
|
|
|
#else /* LTC_NO_ROLC */
|
|
|
|
#define ROL64c ROL64
|
|
#define ROR64c ROR64
|
|
|
|
#endif
|
|
|
|
#else /* Not x86_64 */
|
|
|
|
#define ROL64(x, y) \
|
|
( (((x)<<((ulong64)(y)&63)) | \
|
|
(((x)&CONST64(0xFFFFFFFFFFFFFFFF))>>((ulong64)64-((y)&63)))) & CONST64(0xFFFFFFFFFFFFFFFF))
|
|
|
|
#define ROR64(x, y) \
|
|
( ((((x)&CONST64(0xFFFFFFFFFFFFFFFF))>>((ulong64)(y)&CONST64(63))) | \
|
|
((x)<<((ulong64)(64-((y)&CONST64(63)))))) & CONST64(0xFFFFFFFFFFFFFFFF))
|
|
|
|
#define ROL64c(x, y) \
|
|
( (((x)<<((ulong64)(y)&63)) | \
|
|
(((x)&CONST64(0xFFFFFFFFFFFFFFFF))>>((ulong64)64-((y)&63)))) & CONST64(0xFFFFFFFFFFFFFFFF))
|
|
|
|
#define ROR64c(x, y) \
|
|
( ((((x)&CONST64(0xFFFFFFFFFFFFFFFF))>>((ulong64)(y)&CONST64(63))) | \
|
|
((x)<<((ulong64)(64-((y)&CONST64(63)))))) & CONST64(0xFFFFFFFFFFFFFFFF))
|
|
|
|
#endif
|
|
|
|
#ifndef MAX
|
|
#define MAX(x, y) ( ((x)>(y))?(x):(y) )
|
|
#endif
|
|
|
|
#ifndef MIN
|
|
#define MIN(x, y) ( ((x)<(y))?(x):(y) )
|
|
#endif
|
|
|
|
/* extract a byte portably */
|
|
#ifdef _MSC_VER
|
|
#define byte(x, n) ((unsigned char)((x) >> (8 * (n))))
|
|
#else
|
|
#define byte(x, n) (((x) >> (8 * (n))) & 255)
|
|
#endif
|
|
|
|
/* $Source$ */
|
|
/* $Revision$ */
|
|
/* $Date$ */
|