mirror of
git://git.musl-libc.org/musl
synced 2024-12-16 19:55:38 +00:00
c4dd0c98ba
aside from the obvious C++ ABI purpose for this change, it also brings musl into alignment with the compiler's idea of the definition of wint_t (use in -Wformat), and makes the situation less awkward on ARM, where wchar_t is unsigned. internal code using wint_t and WEOF was checked against this change, and while a few cases of storing WEOF into wchar_t were found, they all seem to operate properly with the natural conversion from unsigned to signed.
118 lines
2.5 KiB
C
118 lines
2.5 KiB
C
#ifndef _STDINT_H
|
|
#define _STDINT_H
|
|
|
|
#define __NEED_int8_t
|
|
#define __NEED_int16_t
|
|
#define __NEED_int32_t
|
|
#define __NEED_int64_t
|
|
|
|
#define __NEED_uint8_t
|
|
#define __NEED_uint16_t
|
|
#define __NEED_uint32_t
|
|
#define __NEED_uint64_t
|
|
|
|
#define __NEED_intptr_t
|
|
#define __NEED_uintptr_t
|
|
|
|
#define __NEED_intmax_t
|
|
#define __NEED_uintmax_t
|
|
|
|
#include <bits/alltypes.h>
|
|
|
|
typedef int8_t int_fast8_t;
|
|
typedef int64_t int_fast64_t;
|
|
|
|
typedef int8_t int_least8_t;
|
|
typedef int16_t int_least16_t;
|
|
typedef int32_t int_least32_t;
|
|
typedef int64_t int_least64_t;
|
|
|
|
typedef uint8_t uint_fast8_t;
|
|
typedef uint64_t uint_fast64_t;
|
|
|
|
typedef uint8_t uint_least8_t;
|
|
typedef uint16_t uint_least16_t;
|
|
typedef uint32_t uint_least32_t;
|
|
typedef uint64_t uint_least64_t;
|
|
|
|
#define INT8_MIN (-1-0x7f)
|
|
#define INT16_MIN (-1-0x7fff)
|
|
#define INT32_MIN (-1-0x7fffffff)
|
|
#define INT64_MIN (-1-0x7fffffffffffffff)
|
|
|
|
#define INT8_MAX (0x7f)
|
|
#define INT16_MAX (0x7fff)
|
|
#define INT32_MAX (0x7fffffff)
|
|
#define INT64_MAX (0x7fffffffffffffff)
|
|
|
|
#define UINT8_MAX (0xff)
|
|
#define UINT16_MAX (0xffff)
|
|
#define UINT32_MAX (0xffffffff)
|
|
#define UINT64_MAX (0xffffffffffffffff)
|
|
|
|
#define INT_FAST8_MIN INT8_MIN
|
|
#define INT_FAST64_MIN INT64_MIN
|
|
|
|
#define INT_LEAST8_MIN INT8_MIN
|
|
#define INT_LEAST16_MIN INT16_MIN
|
|
#define INT_LEAST32_MIN INT32_MIN
|
|
#define INT_LEAST64_MIN INT64_MIN
|
|
|
|
#define INT_FAST8_MAX INT8_MAX
|
|
#define INT_FAST64_MAX INT64_MAX
|
|
|
|
#define INT_LEAST8_MAX INT8_MAX
|
|
#define INT_LEAST16_MAX INT16_MAX
|
|
#define INT_LEAST32_MAX INT32_MAX
|
|
#define INT_LEAST64_MAX INT64_MAX
|
|
|
|
#define UINT_FAST8_MAX UINT8_MAX
|
|
#define UINT_FAST64_MAX UINT64_MAX
|
|
|
|
#define UINT_LEAST8_MAX UINT8_MAX
|
|
#define UINT_LEAST16_MAX UINT16_MAX
|
|
#define UINT_LEAST32_MAX UINT32_MAX
|
|
#define UINT_LEAST64_MAX UINT64_MAX
|
|
|
|
#define INTMAX_MIN INT64_MIN
|
|
#define INTMAX_MAX INT64_MAX
|
|
#define UINTMAX_MAX UINT64_MAX
|
|
|
|
#define WINT_MIN 0
|
|
#define WINT_MAX UINT32_MAX
|
|
|
|
#if L'\0'-1 > 0
|
|
#define WCHAR_MAX (0xffffffffu+L'\0')
|
|
#define WCHAR_MIN (0+L'\0')
|
|
#else
|
|
#define WCHAR_MAX (0x7fffffff+L'\0')
|
|
#define WCHAR_MIN (-1-0x7fffffff+L'\0')
|
|
#endif
|
|
|
|
#define SIG_ATOMIC_MIN INT32_MIN
|
|
#define SIG_ATOMIC_MAX INT32_MAX
|
|
|
|
#include <bits/stdint.h>
|
|
|
|
#define INT8_C(c) c
|
|
#define INT16_C(c) c
|
|
#define INT32_C(c) c
|
|
|
|
#define UINT8_C(c) c
|
|
#define UINT16_C(c) c
|
|
#define UINT32_C(c) c ## U
|
|
|
|
#if UINTPTR_MAX == UINT64_MAX
|
|
#define INT64_C(c) c ## L
|
|
#define UINT64_C(c) c ## UL
|
|
#define INTMAX_C(c) c ## L
|
|
#define UINTMAX_C(c) c ## UL
|
|
#else
|
|
#define INT64_C(c) c ## LL
|
|
#define UINT64_C(c) c ## ULL
|
|
#define INTMAX_C(c) c ## LL
|
|
#define UINTMAX_C(c) c ## ULL
|
|
#endif
|
|
|
|
#endif
|