#include <limits.h>
#include <stddef.h>
-#if __STDC_VERSION__ >= 199900l
+#if __STDC_VERSION__ >= 199901
# include <stdint.h>
#endif
+#ifndef MLIB_COMPILER_H
+# include "compiler.h"
+#endif
+
/*----- Decide on some types ----------------------------------------------*/
/* --- Make GNU C shut up --- */
/* --- List macros --- */
#ifdef HAVE_UINT64
-# define DOUINTCONV(_) \
- _(8, 8, 8) \
- _(16, 16, 16) _(16, 16_L, 16l) _(16, 16_B, 16b) \
- _(24, 24, 24) _(24, 24_L, 24l) _(24, 24_B, 24b) \
- _(32, 32, 32) _(32, 32_L, 32l) _(32, 32_B, 32b) \
+# define DOUINTCONV_64(_) \
_(64, 64, 64) _(64, 64_L, 64l) _(64, 64_B, 64b)
-# define DOUINTSZ(_) _(8) _(16) _(24) _(32) _(64)
+# define DOUINTSZ_64(_) _(64)
#else
-# define DOUINTCONV(_) \
+# define DOUINTCONV_64(_)
+# define DOUINTSZ_64(_)
+#endif
+
+#define DOUINTCONV(_) \
_(8, 8, 8) \
_(16, 16, 16) _(16, 16_L, 16l) _(16, 16_B, 16b) \
_(24, 24, 24) _(24, 24_L, 24l) _(24, 24_B, 24b) \
- _(32, 32, 32) _(32, 32_L, 32l) _(32, 32_B, 32b)
-# define DOUINTSZ(_) _(8) _(16) _(24) _(32)
-#endif
+ _(32, 32, 32) _(32, 32_L, 32l) _(32, 32_B, 32b) \
+ DOUINTCONV_64(_)
+#define DOUINTSZ(_) _(8) _(16) _(24) _(32) _DOUINTSZ_64(_)
/* --- Type coercions --- */
-#define U8(x) ((octet)((x) & MASK8))
-#define U16(x) ((uint16)((x) & MASK16))
-#define U24(x) ((uint24)((x) & MASK24))
-#define U32(x) ((uint32)((x) & MASK32))
+#define U8(x) ((octet)((x)&MASK8))
+#define U16(x) ((uint16)((x)&MASK16))
+#define U24(x) ((uint24)((x)&MASK24))
+#define U32(x) ((uint32)((x)&MASK32))
#ifdef HAVE_UINT64
-# define U64(x) ((uint64)(x) & MASK64)
+# define U64(x) ((uint64)(x)&MASK64)
# define U64_(d, x) ((d).i = U64(x).i)
#else
# define U64_(d, x) ((d).hi = U32((x).hi), (d).lo = U32((x).lo))
/* --- Safe shifting macros --- */
-#define LSL8(v, s) U8(U8(v) << ((s) & 7u))
-#define LSR8(v, s) U8(U8(v) >> ((s) & 7u))
-#define LSL16(v, s) U16(U16(v) << ((s) & 15u))
-#define LSR16(v, s) U16(U16(v) >> ((s) & 15u))
-#define LSL24(v, s) U24(U24(v) << ((s) % 24u))
-#define LSR24(v, s) U24(U24(v) >> ((s) % 24u))
-#define LSL32(v, s) U32(U32(v) << ((s) & 31u))
-#define LSR32(v, s) U32(U32(v) >> ((s) & 31u))
+#define LSL8(v, s) U8(U8(v) << ((s)&7u))
+#define LSR8(v, s) U8(U8(v) >> ((s)&7u))
+#define LSL16(v, s) U16(U16(v) << ((s)&15u))
+#define LSR16(v, s) U16(U16(v) >> ((s)&15u))
+#define LSL24(v, s) U24(U24(v) << ((s)%24u))
+#define LSR24(v, s) U24(U24(v) >> ((s)%24u))
+#define LSL32(v, s) U32(U32(v) << ((s)&31u))
+#define LSR32(v, s) U32(U32(v) >> ((s)&31u))
#ifdef HAVE_UINT64
-# define LSL64(v, s) U64(U64(v) << ((s) & 63u))
-# define LSR64(v, s) U64(U64(v) >> ((s) & 63u))
+# define LSL64(v, s) U64(U64(v) << ((s)&63u))
+# define LSR64(v, s) U64(U64(v) >> ((s)&63u))
# define LSL64_(d, v, s) ((d).i = LSL64((v).i, (s)))
# define LSR64_(d, v, s) ((d).i = LSR64((v).i, (s)))
#else
# define LSL64_(d, v, s) do { \
- unsigned _s = (s) & 63u; \
+ unsigned _s = (s)&63u; \
uint32 _l = (v).lo, _h = (v).hi; \
kludge64 *_d = &(d); \
if (_s >= 32) { \
} \
} while (0)
# define LSR64_(d, v, s) do { \
- unsigned _s = (s) & 63u; \
+ unsigned _s = (s)&63u; \
uint32 _l = (v).lo, _h = (v).hi; \
kludge64 *_d = &(d); \
if (_s >= 32) { \
# define ROR64_(d, v, s) ((d).i = ROR64((v).i, (s)))
#else
# define ROL64_(d, v, s) do { \
- unsigned _s = (s) & 63u; \
+ unsigned _s = (s)&63u; \
uint32 _l = (v).lo, _h = (v).hi; \
kludge64 *_d = &(d); \
if (_s > 32) { \
} \
} while (0)
# define ROR64_(d, v, s) do { \
- unsigned _s = (s) & 63u; \
+ unsigned _s = (s)&63u; \
uint32 _l = (v).lo, _h = (v).hi; \
kludge64 *_d = &(d); \
if (_s > 32) { \
} while (0)
#endif
+/* --- Endianness swapping --- */
+
+#if GCC_VERSION_P(4, 8) || CLANG_VERSION_P(3, 2)
+# define ENDSWAP16(x) ((uint16)__builtin_bswap16(x))
+#endif
+#if GCC_VERSION_P(4, 3) || CLANG_VERSION_P(3, 2)
+# define ENDSWAP32(x) ((uint32)__builtin_bswap32(x))
+#endif
+#if (GCC_VERSION_P(4, 3) || CLANG_VERSION_P(3, 2)) && defined(HAVE_UINT64)
+# define ENDSWAP64(x) ((uint64)__builtin_bswap64(x))
+#endif
+
+#ifndef ENDSWAP8
+# define ENDSWAP8(x) U8(x)
+#endif
+#ifndef ENDSWAP16
+# define ENDSWAP16(x) \
+ ((((uint16)(x) >> 8)&0xff) | \
+ (((uint16)(x)&0xff) << 8))
+#endif
+#ifndef ENDSWAP24
+# define ENDSWAP24(x) \
+ ((((uint24)(x) >> 16)&0xff) | \
+ ((uint24)(x)&0xff00) | \
+ ((uint24)((x)&0xff) << 16))
+#endif
+#ifndef ENDSWAP32
+# define ENDSWAP32(x) \
+ (ENDSWAP16(((uint32)(x) >> 16)&0xffff) | \
+ ((uint32)ENDSWAP16((x)&0xffff) << 16))
+#endif
+#if defined(HAVE_UINT64) && !defined(ENDSWAP64)
+# define ENDSWAP64(x) \
+ (ENDSWAP32(((uint64)(x) >> 32)&0xffffffff) | \
+ ((uint64)ENDSWAP32((x)&0xffffffff) << 32))
+#endif
+#ifdef HAVE_UINT64
+# define ENDSWAP64_(z, x) \
+ ((z).i = ENDSWAP64((x).i))
+#else
+# define ENDSWAP64_(z, x) \
+ ((z).lo = ENDSWAP32((x).hi), \
+ (z).hi = ENDSWAP32((x).lo))
+#endif
+
+#define MLIB_LITTLE_ENDIAN 1234
+#define MLIB_BIG_ENDIAN 4321
+#if defined(__ORDER_LITTLE_ENDIAN__) && \
+ __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
+# define MLIB_BYTE_ORDER MLIB_LITTLE_ENDIAN
+#elif defined(__ORDER_BIG_ENDIAN__) && \
+ __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
+# define MLIB_BYTE_ORDER MLIB_BIG_ENDIAN
+#endif
+
+#if MLIB_BYTE_ORDER == MLIB_LITTLE_ENDIAN
+# define HTOL16(x) (x)
+# define LTOH16(x) (x)
+# define HTOB16(x) ENDSWAP16(x)
+# define BTOH16(x) ENDSWAP16(x)
+# define HTOL24(x) (x)
+# define LTOH24(x) (x)
+# define HTOB24(x) ENDSWAP24(x)
+# define BTOH24(x) ENDSWAP24(x)
+# define HTOL32(x) (x)
+# define LTOH32(x) (x)
+# define HTOB32(x) ENDSWAP32(x)
+# define BTOH32(x) ENDSWAP32(x)
+# ifdef HAVE_UINT64
+# define HTOL64(x) (x)
+# define LTOH64(x) (x)
+# define HTOB64(x) ENDSWAP64(x)
+# define BTOH64(x) ENDSWAP64(x)
+# endif
+# define HTOL64_(z, x) ASSIGN64(z, x)
+# define LTOH64_(z, x) ASSIGN64(z, x)
+# define HTOB64_(z, x) ENDSWAP64_(z, x)
+# define BTOH64_(z, x) ENDSWAP64_(z, x)
+#elif MLIB_BYTE_ORDER == MLIB_BIG_ENDIAN
+# define HTOL16(x) ENDSWAP16(x)
+# define LTOH16(x) ENDSWAP16(x)
+# define HTOB16(x) (x)
+# define BTOH16(x) (x)
+# define HTOL24(x) ENDSWAP24(x)
+# define LTOH24(x) ENDSWAP24(x)
+# define HTOB24(x) (x)
+# define BTOH24(x) (x)
+# define HTOL32(x) ENDSWAP32(x)
+# define LTOH32(x) ENDSWAP32(x)
+# define HTOB32(x) (x)
+# define BTOH32(x) (x)
+# ifdef HAVE_UINT64
+# define HTOL64(x) ENDSWAP64(x)
+# define LTOH64(x) ENDSWAP64(x)
+# define HTOB64(x) (x)
+# define BTOH64(x) (x)
+# endif
+# define HTOL64_(z, x) ENDSWAP64_(z, x)
+# define LTOH64_(z, x) ENDSWAP64_(z, x)
+# define HTOB64_(z, x) ASSIGN64(z, x)
+# define BTOH64_(z, x) ASSIGN64(z, x)
+#endif
+
+/* --- Unaligned access (GCC-specific) --- */
+
+#if (GCC_VERSION_P(3, 3) || CLANG_VERSION_P(3, 0)) && CHAR_BIT == 8
+# define MLIB_MISALIGNED __attribute__((aligned(1), may_alias))
+# if __SIZEOF_SHORT__ == 2
+ typedef MLIB_MISALIGNED unsigned short misaligned_uint16;
+# define RAW16(p) (*(misaligned_uint16 *)(p))
+# endif
+# if __SIZEOF_INT__ == 4
+ typedef MLIB_MISALIGNED unsigned int misaligned_uint32;
+# define RAW32(p) (*(misaligned_uint32 *)(p))
+# elif __SIZEOF_LONG__ == 4
+ typedef MLIB_MISALIGNED unsigned long misaligned_uint32;
+# define RAW32(p) (*(misaligned_uint32 *)(p))
+# endif
+# if __SIZEOF_LONG__ == 8
+ typedef MLIB_MISALIGNED unsigned long misaligned_uint64;
+# define RAW64(p) (*(misaligned_uint64 *)(p))
+# elif __SIZEOF_LONG_LONG__ == 8
+ typedef MLIB_MISALIGNED unsigned long long misaligned_uint64;
+# define RAW64(p) (*(misaligned_uint64 *)(p))
+# endif
+#endif
+
/* --- Storage and retrieval --- */
-#define GETBYTE(p, o) (((octet *)(p))[o] & MASK8)
+#if defined(RAW16) && defined(LTOH16)
+# define LOAD16_L(p) LTOH16(RAW16(p))
+#endif
+#if defined(RAW16) && defined(HTOL16)
+# define STORE16_L(p, x) (RAW16(p) = HTOL16(x))
+#endif
+#if defined(RAW16) && defined(BTOH16)
+# define LOAD16_B(p) BTOH16(RAW16(p))
+#endif
+#if defined(RAW16) && defined(HTOB16)
+# define STORE16_B(p, x) (RAW16(p) = HTOB16(x))
+#endif
+
+#if defined(RAW32) && defined(LTOH32)
+# define LOAD32_L(p) LTOH32(RAW32(p))
+#endif
+#if defined(RAW32) && defined(HTOL32)
+# define STORE32_L(p, x) (RAW32(p) = HTOL32(x))
+#endif
+#if defined(RAW32) && defined(BTOH32)
+# define LOAD32_B(p) BTOH32(RAW32(p))
+#endif
+#if defined(RAW32) && defined(HTOB32)
+# define STORE32_B(p, x) (RAW32(p) = HTOB32(x))
+#endif
+
+#if defined(RAW64) && defined(LTOH64)
+# define LOAD64_L(p) LTOH64(RAW64(p))
+#endif
+#if defined(RAW64) && defined(HTOL64)
+# define STORE64_L(p, x) (RAW64(p) = HTOL64(x))
+#endif
+#if defined(RAW64) && defined(BTOH64)
+# define LOAD64_B(p) BTOH64(RAW64(p))
+#endif
+#if defined(RAW64) && defined(HTOB64)
+# define STORE64_B(p, x) (RAW64(p) = HTOB64(x))
+#endif
+
+#define GETBYTE(p, o) (((octet *)(p))[o]&MASK8)
#define PUTBYTE(p, o, v) (((octet *)(p))[o] = U8((v)))
#define LOAD8(p) (GETBYTE((p), 0))
#else
# define LOAD64_B_(d, p) \
- ((d).hi = LOAD32_B((octet *)(p) + 0), \
- (d).lo = LOAD32_B((octet *)(p) + 4))
+ ((d).hi = LOAD32_B((octet *)(p) + 0), \
+ (d).lo = LOAD32_B((octet *)(p) + 4))
# define LOAD64_L_(d, p) \
- ((d).lo = LOAD32_L((octet *)(p) + 0), \
- (d).hi = LOAD32_L((octet *)(p) + 4))
+ ((d).lo = LOAD32_L((octet *)(p) + 0), \
+ (d).hi = LOAD32_L((octet *)(p) + 4))
# define LOAD64_(d, p) LOAD64_B_((d), (p))
# define STORE64_B_(p, v) \
- (STORE32_B((octet *)(p) + 0, (v).hi), \
- STORE32_B((octet *)(p) + 4, (v).lo))
+ (STORE32_B((octet *)(p) + 0, (v).hi), \
+ STORE32_B((octet *)(p) + 4, (v).lo))
# define STORE64_L_(p, v) \
- (STORE32_L((octet *)(p) + 0, (v).lo), \
- STORE32_L((octet *)(p) + 4, (v).hi))
+ (STORE32_L((octet *)(p) + 0, (v).lo), \
+ STORE32_L((octet *)(p) + 4, (v).hi))
# define STORE64_(p, v) STORE64_B_((p), (v))
#endif
/* --- Other operations on 64-bit integers --- */
#ifdef HAVE_UINT64
-# define SET64(d, h, l) ((d).i = (U64((h)) << 32) | U64((l)))
+# define SET64(d, h, l) ((d).i = ((uint64)(U32(h)) << 32) | U32(l))
# define ASSIGN64(d, x) ((d).i = U64((x)))
# define HI64(x) U32((x).i >> 32)
# define LO64(x) U32((x).i)
# define GET64(t, x) ((t)(x).i)
+# define SETBYTE64(z, x, j) ((z).i |= (uint64)U8(x) << 8*(j))
#else
# define SET64(d, h, l) ((d).hi = U32(h), (d).lo = U32(l))
# define ASSIGN64(d, x) \
- ((d).hi = ((x & ~MASK32) >> 16) >> 16, (d).lo = U32(x))
+ ((d).hi = ((x&~MASK32) >> 16) >> 16, (d).lo = U32(x))
# define HI64(x) U32((x).hi)
# define LO64(x) U32((x).lo)
-# define GET64(t, x) (((((t)HI64(x) << 16) << 16) & ~MASK32) | (t)LO64(x))
+# define GET64(t, x) (((((t)HI64(x) << 16) << 16)&~MASK32) | (t)LO64(x))
+# define SETBYTE64(z, x, j) \
+ ((j) < 4 ? (z).lo |= (uint32)U8(x) << 8*(j) \
+ : (z).hi |= (uint32)U8(x) << 8*((j) - 4))
#endif
#ifdef HAVE_UINT64
-# define AND64(d, x, y) ((d).i = (x).i & (y).i)
+# define AND64(d, x, y) ((d).i = (x).i&(y).i)
# define OR64(d, x, y) ((d).i = (x).i | (y).i)
# define XOR64(d, x, y) ((d).i = (x).i ^ (y).i)
# define CPL64(d, x) ((d).i = ~(x).i)
# define ADD64(d, x, y) ((d).i = (x).i + (y).i)
# define SUB64(d, x, y) ((d).i = (x).i - (y).i)
# define CMP64(x, op, y) ((x).i op (y).i)
-# define ZERO64(x) ((x) == 0)
+# define ZERO64(x) ((x).i == 0)
#else
-# define AND64(d, x, y) ((d).lo = (x).lo & (y).lo, (d).hi = (x).hi & (y).hi)
+# define AND64(d, x, y) ((d).lo = (x).lo&(y).lo, (d).hi = (x).hi&(y).hi)
# define OR64(d, x, y) ((d).lo = (x).lo | (y).lo, (d).hi = (x).hi | (y).hi)
# define XOR64(d, x, y) ((d).lo = (x).lo ^ (y).lo, (d).hi = (x).hi ^ (y).hi)
# define CPL64(d, x) ((d).lo = ~(x).lo, (d).hi = ~(x).hi)