Home
last modified time | relevance | path

Searched refs:__force (Results 1 – 25 of 906) sorted by relevance

12345678910>>...37

/linux-6.1.9/include/scsi/
Dscsi_devinfo.h9 #define BLIST_NOLUN ((__force blist_flags_t)(1ULL << 0))
12 #define BLIST_FORCELUN ((__force blist_flags_t)(1ULL << 1))
14 #define BLIST_BORKEN ((__force blist_flags_t)(1ULL << 2))
16 #define BLIST_KEY ((__force blist_flags_t)(1ULL << 3))
18 #define BLIST_SINGLELUN ((__force blist_flags_t)(1ULL << 4))
20 #define BLIST_NOTQ ((__force blist_flags_t)(1ULL << 5))
22 #define BLIST_SPARSELUN ((__force blist_flags_t)(1ULL << 6))
24 #define BLIST_MAX5LUN ((__force blist_flags_t)(1ULL << 7))
26 #define BLIST_ISROM ((__force blist_flags_t)(1ULL << 8))
28 #define BLIST_LARGELUN ((__force blist_flags_t)(1ULL << 9))
[all …]
/linux-6.1.9/drivers/net/wireless/intel/iwlwifi/fw/
Dfile.h249 IWL_UCODE_TLV_API_FRAGMENTED_SCAN = (__force iwl_ucode_tlv_api_t)8,
250 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE = (__force iwl_ucode_tlv_api_t)9,
251 IWL_UCODE_TLV_API_LQ_SS_PARAMS = (__force iwl_ucode_tlv_api_t)18,
252 IWL_UCODE_TLV_API_NEW_VERSION = (__force iwl_ucode_tlv_api_t)20,
253 IWL_UCODE_TLV_API_SCAN_TSF_REPORT = (__force iwl_ucode_tlv_api_t)28,
254 IWL_UCODE_TLV_API_TKIP_MIC_KEYS = (__force iwl_ucode_tlv_api_t)29,
255 IWL_UCODE_TLV_API_STA_TYPE = (__force iwl_ucode_tlv_api_t)30,
256 IWL_UCODE_TLV_API_NAN2_VER2 = (__force iwl_ucode_tlv_api_t)31,
258 IWL_UCODE_TLV_API_ADAPTIVE_DWELL = (__force iwl_ucode_tlv_api_t)32,
259 IWL_UCODE_TLV_API_OCE = (__force iwl_ucode_tlv_api_t)33,
[all …]
/linux-6.1.9/include/linux/
Dblk_types.h94 #define BLK_STS_NOTSUPP ((__force blk_status_t)1)
95 #define BLK_STS_TIMEOUT ((__force blk_status_t)2)
96 #define BLK_STS_NOSPC ((__force blk_status_t)3)
97 #define BLK_STS_TRANSPORT ((__force blk_status_t)4)
98 #define BLK_STS_TARGET ((__force blk_status_t)5)
99 #define BLK_STS_NEXUS ((__force blk_status_t)6)
100 #define BLK_STS_MEDIUM ((__force blk_status_t)7)
101 #define BLK_STS_PROTECTION ((__force blk_status_t)8)
102 #define BLK_STS_RESOURCE ((__force blk_status_t)9)
103 #define BLK_STS_IOERR ((__force blk_status_t)10)
[all …]
Dgfp_types.h72 #define __GFP_DMA ((__force gfp_t)___GFP_DMA)
73 #define __GFP_HIGHMEM ((__force gfp_t)___GFP_HIGHMEM)
74 #define __GFP_DMA32 ((__force gfp_t)___GFP_DMA32)
75 #define __GFP_MOVABLE ((__force gfp_t)___GFP_MOVABLE) /* ZONE_MOVABLE allowed */
105 #define __GFP_RECLAIMABLE ((__force gfp_t)___GFP_RECLAIMABLE)
106 #define __GFP_WRITE ((__force gfp_t)___GFP_WRITE)
107 #define __GFP_HARDWALL ((__force gfp_t)___GFP_HARDWALL)
108 #define __GFP_THISNODE ((__force gfp_t)___GFP_THISNODE)
109 #define __GFP_ACCOUNT ((__force gfp_t)___GFP_ACCOUNT)
138 #define __GFP_ATOMIC ((__force gfp_t)___GFP_ATOMIC)
[all …]
Dvirtio_byteorder.h19 return le16_to_cpu((__force __le16)val); in __virtio16_to_cpu()
21 return be16_to_cpu((__force __be16)val); in __virtio16_to_cpu()
27 return (__force __virtio16)cpu_to_le16(val); in __cpu_to_virtio16()
29 return (__force __virtio16)cpu_to_be16(val); in __cpu_to_virtio16()
35 return le32_to_cpu((__force __le32)val); in __virtio32_to_cpu()
37 return be32_to_cpu((__force __be32)val); in __virtio32_to_cpu()
43 return (__force __virtio32)cpu_to_le32(val); in __cpu_to_virtio32()
45 return (__force __virtio32)cpu_to_be32(val); in __cpu_to_virtio32()
51 return le64_to_cpu((__force __le64)val); in __virtio64_to_cpu()
53 return be64_to_cpu((__force __be64)val); in __virtio64_to_cpu()
[all …]
Dserial_core.h501 #define UPF_FOURPORT ((__force upf_t) ASYNC_FOURPORT /* 1 */ )
502 #define UPF_SAK ((__force upf_t) ASYNC_SAK /* 2 */ )
503 #define UPF_SPD_HI ((__force upf_t) ASYNC_SPD_HI /* 4 */ )
504 #define UPF_SPD_VHI ((__force upf_t) ASYNC_SPD_VHI /* 5 */ )
505 #define UPF_SPD_CUST ((__force upf_t) ASYNC_SPD_CUST /* 0x0030 */ )
506 #define UPF_SPD_WARP ((__force upf_t) ASYNC_SPD_WARP /* 0x1010 */ )
507 #define UPF_SPD_MASK ((__force upf_t) ASYNC_SPD_MASK /* 0x1030 */ )
508 #define UPF_SKIP_TEST ((__force upf_t) ASYNC_SKIP_TEST /* 6 */ )
509 #define UPF_AUTO_IRQ ((__force upf_t) ASYNC_AUTO_IRQ /* 7 */ )
510 #define UPF_HARDPPS_CD ((__force upf_t) ASYNC_HARDPPS_CD /* 11 */ )
[all …]
/linux-6.1.9/include/uapi/linux/byteorder/
Dbig_endian.h16 #define __constant_htonl(x) ((__force __be32)(__u32)(x))
17 #define __constant_ntohl(x) ((__force __u32)(__be32)(x))
18 #define __constant_htons(x) ((__force __be16)(__u16)(x))
19 #define __constant_ntohs(x) ((__force __u16)(__be16)(x))
20 #define __constant_cpu_to_le64(x) ((__force __le64)___constant_swab64((x)))
21 #define __constant_le64_to_cpu(x) ___constant_swab64((__force __u64)(__le64)(x))
22 #define __constant_cpu_to_le32(x) ((__force __le32)___constant_swab32((x)))
23 #define __constant_le32_to_cpu(x) ___constant_swab32((__force __u32)(__le32)(x))
24 #define __constant_cpu_to_le16(x) ((__force __le16)___constant_swab16((x)))
25 #define __constant_le16_to_cpu(x) ___constant_swab16((__force __u16)(__le16)(x))
[all …]
Dlittle_endian.h16 #define __constant_htonl(x) ((__force __be32)___constant_swab32((x)))
17 #define __constant_ntohl(x) ___constant_swab32((__force __be32)(x))
18 #define __constant_htons(x) ((__force __be16)___constant_swab16((x)))
19 #define __constant_ntohs(x) ___constant_swab16((__force __be16)(x))
20 #define __constant_cpu_to_le64(x) ((__force __le64)(__u64)(x))
21 #define __constant_le64_to_cpu(x) ((__force __u64)(__le64)(x))
22 #define __constant_cpu_to_le32(x) ((__force __le32)(__u32)(x))
23 #define __constant_le32_to_cpu(x) ((__force __u32)(__le32)(x))
24 #define __constant_cpu_to_le16(x) ((__force __le16)(__u16)(x))
25 #define __constant_le16_to_cpu(x) ((__force __u16)(__le16)(x))
[all …]
/linux-6.1.9/include/uapi/linux/
Deventpoll.h31 #define EPOLLIN (__force __poll_t)0x00000001
32 #define EPOLLPRI (__force __poll_t)0x00000002
33 #define EPOLLOUT (__force __poll_t)0x00000004
34 #define EPOLLERR (__force __poll_t)0x00000008
35 #define EPOLLHUP (__force __poll_t)0x00000010
36 #define EPOLLNVAL (__force __poll_t)0x00000020
37 #define EPOLLRDNORM (__force __poll_t)0x00000040
38 #define EPOLLRDBAND (__force __poll_t)0x00000080
39 #define EPOLLWRNORM (__force __poll_t)0x00000100
40 #define EPOLLWRBAND (__force __poll_t)0x00000200
[all …]
/linux-6.1.9/tools/include/uapi/sound/
Dasound.h173 #define SNDRV_PCM_ACCESS_MMAP_INTERLEAVED ((__force snd_pcm_access_t) 0) /* interleaved mmap */
174 #define SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED ((__force snd_pcm_access_t) 1) /* noninterleaved mmap …
175 #define SNDRV_PCM_ACCESS_MMAP_COMPLEX ((__force snd_pcm_access_t) 2) /* complex mmap */
176 #define SNDRV_PCM_ACCESS_RW_INTERLEAVED ((__force snd_pcm_access_t) 3) /* readi/writei */
177 #define SNDRV_PCM_ACCESS_RW_NONINTERLEAVED ((__force snd_pcm_access_t) 4) /* readn/writen */
181 #define SNDRV_PCM_FORMAT_S8 ((__force snd_pcm_format_t) 0)
182 #define SNDRV_PCM_FORMAT_U8 ((__force snd_pcm_format_t) 1)
183 #define SNDRV_PCM_FORMAT_S16_LE ((__force snd_pcm_format_t) 2)
184 #define SNDRV_PCM_FORMAT_S16_BE ((__force snd_pcm_format_t) 3)
185 #define SNDRV_PCM_FORMAT_U16_LE ((__force snd_pcm_format_t) 4)
[all …]
/linux-6.1.9/include/uapi/sound/
Dasound.h173 #define SNDRV_PCM_ACCESS_MMAP_INTERLEAVED ((__force snd_pcm_access_t) 0) /* interleaved mmap */
174 #define SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED ((__force snd_pcm_access_t) 1) /* noninterleaved mmap …
175 #define SNDRV_PCM_ACCESS_MMAP_COMPLEX ((__force snd_pcm_access_t) 2) /* complex mmap */
176 #define SNDRV_PCM_ACCESS_RW_INTERLEAVED ((__force snd_pcm_access_t) 3) /* readi/writei */
177 #define SNDRV_PCM_ACCESS_RW_NONINTERLEAVED ((__force snd_pcm_access_t) 4) /* readn/writen */
181 #define SNDRV_PCM_FORMAT_S8 ((__force snd_pcm_format_t) 0)
182 #define SNDRV_PCM_FORMAT_U8 ((__force snd_pcm_format_t) 1)
183 #define SNDRV_PCM_FORMAT_S16_LE ((__force snd_pcm_format_t) 2)
184 #define SNDRV_PCM_FORMAT_S16_BE ((__force snd_pcm_format_t) 3)
185 #define SNDRV_PCM_FORMAT_U16_LE ((__force snd_pcm_format_t) 4)
[all …]
/linux-6.1.9/net/ipv6/
Dip6_checksum.c16 __u32 sum = (__force u32)csum; in csum_ipv6_magic()
18 sum += (__force u32)saddr->s6_addr32[0]; in csum_ipv6_magic()
19 carry = (sum < (__force u32)saddr->s6_addr32[0]); in csum_ipv6_magic()
22 sum += (__force u32)saddr->s6_addr32[1]; in csum_ipv6_magic()
23 carry = (sum < (__force u32)saddr->s6_addr32[1]); in csum_ipv6_magic()
26 sum += (__force u32)saddr->s6_addr32[2]; in csum_ipv6_magic()
27 carry = (sum < (__force u32)saddr->s6_addr32[2]); in csum_ipv6_magic()
30 sum += (__force u32)saddr->s6_addr32[3]; in csum_ipv6_magic()
31 carry = (sum < (__force u32)saddr->s6_addr32[3]); in csum_ipv6_magic()
34 sum += (__force u32)daddr->s6_addr32[0]; in csum_ipv6_magic()
[all …]
/linux-6.1.9/arch/s390/include/asm/
Dchecksum.h49 u32 csum = (__force u32) sum; in csum_fold()
53 return (__force __sum16) ~csum; in csum_fold()
73 return csum_fold((__force __wsum)(csum >> 32)); in ip_fast_csum()
83 __u64 csum = (__force __u64)sum; in csum_tcpudp_nofold()
85 csum += (__force __u32)saddr; in csum_tcpudp_nofold()
86 csum += (__force __u32)daddr; in csum_tcpudp_nofold()
90 return (__force __wsum)(csum >> 32); in csum_tcpudp_nofold()
116 __u64 sum = (__force __u64)csum; in csum_ipv6_magic()
118 sum += (__force __u32)saddr->s6_addr32[0]; in csum_ipv6_magic()
119 sum += (__force __u32)saddr->s6_addr32[1]; in csum_ipv6_magic()
[all …]
/linux-6.1.9/include/linux/rpmsg/
Dbyteorder.h22 return le16_to_cpu((__force __le16)val); in __rpmsg16_to_cpu()
24 return be16_to_cpu((__force __be16)val); in __rpmsg16_to_cpu()
30 return (__force __rpmsg16)cpu_to_le16(val); in __cpu_to_rpmsg16()
32 return (__force __rpmsg16)cpu_to_be16(val); in __cpu_to_rpmsg16()
38 return le32_to_cpu((__force __le32)val); in __rpmsg32_to_cpu()
40 return be32_to_cpu((__force __be32)val); in __rpmsg32_to_cpu()
46 return (__force __rpmsg32)cpu_to_le32(val); in __cpu_to_rpmsg32()
48 return (__force __rpmsg32)cpu_to_be32(val); in __cpu_to_rpmsg32()
54 return le64_to_cpu((__force __le64)val); in __rpmsg64_to_cpu()
56 return be64_to_cpu((__force __be64)val); in __rpmsg64_to_cpu()
[all …]
/linux-6.1.9/fs/ntfs/
Dendian.h21 return le16_to_cpu((__force le16)x); in sle16_to_cpu()
26 return le32_to_cpu((__force le32)x); in sle32_to_cpu()
31 return le64_to_cpu((__force le64)x); in sle64_to_cpu()
36 return le16_to_cpu(*(__force le16*)x); in sle16_to_cpup()
41 return le32_to_cpu(*(__force le32*)x); in sle32_to_cpup()
46 return le64_to_cpu(*(__force le64*)x); in sle64_to_cpup()
51 return (__force sle16)cpu_to_le16(x); in cpu_to_sle16()
56 return (__force sle32)cpu_to_le32(x); in cpu_to_sle32()
61 return (__force sle64)cpu_to_le64(x); in cpu_to_sle64()
66 return (__force sle16)cpu_to_le16(*x); in cpu_to_sle16p()
[all …]
/linux-6.1.9/arch/powerpc/include/asm/
Dchecksum.h41 u32 tmp = (__force u32)sum; in csum_fold()
49 return (__force __sum16)(~(tmp + rol32(tmp, 16)) >> 16); in csum_fold()
61 u64 s = (__force u32)sum; in csum_tcpudp_nofold()
63 s += (__force u32)saddr; in csum_tcpudp_nofold()
64 s += (__force u32)daddr; in csum_tcpudp_nofold()
70 return (__force __wsum) from64to32(s); in csum_tcpudp_nofold()
98 u64 res = (__force u64)csum; in csum_add()
100 res += (__force u64)addend; in csum_add()
101 return (__force __wsum)((u32)res + (res >> 32)); in csum_add()
119 return (__force __wsum)rol32((__force u32)sum, (offset & 1) << 3); in csum_shift()
[all …]
/linux-6.1.9/include/net/
Dchecksum.h59 u32 res = (__force u32)csum; in csum_add()
60 res += (__force u32)addend; in csum_add()
61 return (__force __wsum)(res + (res < (__force u32)addend)); in csum_add()
72 u16 res = (__force u16)csum; in csum16_add()
74 res += (__force u16)addend; in csum16_add()
75 return (__force __sum16)(res + (res < (__force u16)addend)); in csum16_add()
88 return (__force __wsum)ror32((__force u32)sum, 8); in csum_shift()
113 return (__force __wsum)n; in csum_unfold()
122 #define CSUM_MANGLED_0 ((__force __sum16)0xffff)
131 __wsum tmp = csum_sub(~csum_unfold(*sum), (__force __wsum)from); in csum_replace4()
[all …]
/linux-6.1.9/fs/befs/
Dendian.h19 return le64_to_cpu((__force __le64)n); in fs64_to_cpu()
21 return be64_to_cpu((__force __be64)n); in fs64_to_cpu()
28 return (__force fs64)cpu_to_le64(n); in cpu_to_fs64()
30 return (__force fs64)cpu_to_be64(n); in cpu_to_fs64()
37 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
39 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
46 return (__force fs32)cpu_to_le32(n); in cpu_to_fs32()
48 return (__force fs32)cpu_to_be32(n); in cpu_to_fs32()
55 return le16_to_cpu((__force __le16)n); in fs16_to_cpu()
57 return be16_to_cpu((__force __be16)n); in fs16_to_cpu()
[all …]
/linux-6.1.9/arch/ia64/include/uapi/asm/
Dcmpxchg.h36 __xchg_result = ia64_xchg1((__u8 __force *)ptr, x); \
40 __xchg_result = ia64_xchg2((__u16 __force *)ptr, x); \
44 __xchg_result = ia64_xchg4((__u32 __force *)ptr, x); \
48 __xchg_result = ia64_xchg8((__u64 __force *)ptr, x); \
53 (__typeof__ (*(ptr)) __force) __xchg_result; \
79 _o_ = (__u8) (long __force) (old); \
82 _o_ = (__u16) (long __force) (old); \
85 _o_ = (__u32) (long __force) (old); \
88 _o_ = (__u64) (long __force) (old); \
95 _r_ = ia64_cmpxchg1_##sem((__u8 __force *) ptr, new, _o_); \
[all …]
/linux-6.1.9/include/net/netfilter/
Dnf_queue.h52 if ((__force u32)iph->saddr < (__force u32)iph->daddr) in hash_v4()
53 return jhash_3words((__force u32)iph->saddr, in hash_v4()
54 (__force u32)iph->daddr, iph->protocol, initval); in hash_v4()
56 return jhash_3words((__force u32)iph->daddr, in hash_v4()
57 (__force u32)iph->saddr, iph->protocol, initval); in hash_v4()
64 if ((__force u32)ip6h->saddr.s6_addr32[3] < in hash_v6()
65 (__force u32)ip6h->daddr.s6_addr32[3]) { in hash_v6()
66 a = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6()
67 b = (__force u32) ip6h->daddr.s6_addr32[3]; in hash_v6()
69 b = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6()
[all …]
/linux-6.1.9/arch/alpha/include/asm/
Dio_trivial.h12 return __kernel_ldbu(*(const volatile u8 __force *)a); in IO_CONCAT()
18 return __kernel_ldwu(*(const volatile u16 __force *)a); in IO_CONCAT()
24 __kernel_stb(b, *(volatile u8 __force *)a); in IO_CONCAT()
30 __kernel_stw(b, *(volatile u16 __force *)a); in IO_CONCAT()
38 return *(const volatile u32 __force *)a; in IO_CONCAT()
44 *(volatile u32 __force *)a = b; in IO_CONCAT()
50 return *(const volatile u64 __force *)a; in IO_CONCAT()
56 *(volatile u64 __force *)a = b; in IO_CONCAT()
64 return __kernel_ldbu(*(const volatile u8 __force *)a); in IO_CONCAT()
70 return __kernel_ldwu(*(const volatile u16 __force *)a); in IO_CONCAT()
[all …]
/linux-6.1.9/arch/ia64/lib/
Dchecksum.c41 return (__force __sum16)~from64to16( in csum_tcpudp_magic()
42 (__force u64)saddr + (__force u64)daddr + in csum_tcpudp_magic()
43 (__force u64)sum + ((len + proto) << 8)); in csum_tcpudp_magic()
54 result = (__force u64)saddr + (__force u64)daddr + in csum_tcpudp_nofold()
55 (__force u64)sum + ((len + proto) << 8); in csum_tcpudp_nofold()
62 return (__force __wsum)result; in csum_tcpudp_nofold()
85 result += (__force u32)sum; in csum_partial()
88 return (__force __wsum)result; in csum_partial()
99 return (__force __sum16)~do_csum(buff,len); in ip_compute_csum()
/linux-6.1.9/fs/ufs/
Dswab.h30 return le64_to_cpu((__force __le64)n); in fs64_to_cpu()
32 return be64_to_cpu((__force __be64)n); in fs64_to_cpu()
39 return (__force __fs64)cpu_to_le64(n); in cpu_to_fs64()
41 return (__force __fs64)cpu_to_be64(n); in cpu_to_fs64()
48 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
50 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
57 return (__force __fs32)cpu_to_le32(n); in cpu_to_fs32()
59 return (__force __fs32)cpu_to_be32(n); in cpu_to_fs32()
84 return le16_to_cpu((__force __le16)n); in fs16_to_cpu()
86 return be16_to_cpu((__force __be16)n); in fs16_to_cpu()
[all …]
/linux-6.1.9/arch/alpha/lib/
Dchecksum.c48 return (__force __sum16)~from64to16( in csum_tcpudp_magic()
49 (__force u64)saddr + (__force u64)daddr + in csum_tcpudp_magic()
50 (__force u64)sum + ((len + proto) << 8)); in csum_tcpudp_magic()
59 result = (__force u64)saddr + (__force u64)daddr + in csum_tcpudp_nofold()
60 (__force u64)sum + ((len + proto) << 8); in csum_tcpudp_nofold()
68 return (__force __wsum)result; in csum_tcpudp_nofold()
147 return (__force __sum16)~do_csum(iph,ihl*4); in ip_fast_csum()
168 result += (__force u32)sum; in csum_partial()
171 return (__force __wsum)result; in csum_partial()
182 return (__force __sum16)~from64to16(do_csum(buff,len)); in ip_compute_csum()
/linux-6.1.9/include/asm-generic/
Duaccess.h23 *(u8 *)to = *((u8 __force *)from); in __get_user_fn()
26 *(u16 *)to = get_unaligned((u16 __force *)from); in __get_user_fn()
29 *(u32 *)to = get_unaligned((u32 __force *)from); in __get_user_fn()
32 *(u64 *)to = get_unaligned((u64 __force *)from); in __get_user_fn()
49 *(u8 __force *)to = *(u8 *)from; in __put_user_fn()
52 put_unaligned(*(u16 *)from, (u16 __force *)to); in __put_user_fn()
55 put_unaligned(*(u32 *)from, (u32 __force *)to); in __put_user_fn()
58 put_unaligned(*(u64 *)from, (u64 __force *)to); in __put_user_fn()
84 memcpy(to, (const void __force *)from, n); in raw_copy_from_user()
91 memcpy((void __force *)to, from, n); in raw_copy_to_user()
[all …]

12345678910>>...37