2018-03-21 00:01:05 +03:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
|
|
|
|
#ifndef __ASM_GENERIC_BITS_PER_LONG
|
|
|
|
#define __ASM_GENERIC_BITS_PER_LONG
|
|
|
|
|
2023-07-10 00:23:08 +03:00
|
|
|
#ifndef __BITS_PER_LONG
|
|
|
|
/*
|
|
|
|
* In order to keep safe and avoid regression, only unify uapi
|
|
|
|
* bitsperlong.h for some archs which are using newer toolchains
|
|
|
|
* that have the definitions of __CHAR_BIT__ and __SIZEOF_LONG__.
|
|
|
|
* See the following link for more info:
|
|
|
|
* https://lore.kernel.org/linux-arch/b9624545-2c80-49a1-ac3c-39264a591f7b@app.fastmail.com/
|
|
|
|
*/
|
|
|
|
#if defined(__CHAR_BIT__) && defined(__SIZEOF_LONG__)
|
|
|
|
#define __BITS_PER_LONG (__CHAR_BIT__ * __SIZEOF_LONG__)
|
|
|
|
#else
|
2018-03-21 00:01:05 +03:00
|
|
|
/*
|
|
|
|
* There seems to be no way of detecting this automatically from user
|
|
|
|
* space, so 64 bit architectures should override this in their
|
|
|
|
* bitsperlong.h. In particular, an architecture that supports
|
|
|
|
* both 32 and 64 bit user space must not rely on CONFIG_64BIT
|
|
|
|
* to decide it, but rather check a compiler provided macro.
|
|
|
|
*/
|
|
|
|
#define __BITS_PER_LONG 32
|
|
|
|
#endif
|
2023-07-10 00:23:08 +03:00
|
|
|
#endif
|
2018-03-21 00:01:05 +03:00
|
|
|
|
2024-04-23 12:46:47 +03:00
|
|
|
#ifndef __BITS_PER_LONG_LONG
|
|
|
|
#define __BITS_PER_LONG_LONG 64
|
|
|
|
#endif
|
|
|
|
|
2018-03-21 00:01:05 +03:00
|
|
|
#endif /* __ASM_GENERIC_BITS_PER_LONG */
|