1 /**************************************************************************
3 * Copyright 2008 VMware, Inc.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
41 #if defined(__POPCNT__)
42 #include <popcntintrin.h>
45 #include "c99_compat.h"
53 * Find first bit set in word. Least significant bit is 1.
54 * Return 0 if no bits set.
56 #ifdef HAVE___BUILTIN_FFS
57 #define ffs __builtin_ffs
58 #elif defined(_MSC_VER) && (_M_IX86 || _M_ARM || _M_AMD64 || _M_IA64)
63 if (_BitScanForward(&index
, i
))
73 #ifdef HAVE___BUILTIN_FFSLL
74 #define ffsll __builtin_ffsll
75 #elif defined(_MSC_VER) && (_M_AMD64 || _M_ARM || _M_IA64)
77 ffsll(long long int i
)
80 if (_BitScanForward64(&index
, i
))
87 ffsll(long long int val
);
91 /* Destructively loop over all of the bits in a mask as in:
94 * int i = u_bit_scan(&mymask);
95 * ... process element i
100 u_bit_scan(unsigned *mask
)
102 const int i
= ffs(*mask
) - 1;
108 u_bit_scan64(uint64_t *mask
)
110 const int i
= ffsll(*mask
) - 1;
111 *mask
^= (((uint64_t)1) << i
);
115 /* Determine if an unsigned value is a power of two.
118 * Zero is treated as a power of two.
121 util_is_power_of_two_or_zero(unsigned v
)
123 return (v
& (v
- 1)) == 0;
126 /* Determine if an uint64_t value is a power of two.
129 * Zero is treated as a power of two.
132 util_is_power_of_two_or_zero64(uint64_t v
)
134 return (v
& (v
- 1)) == 0;
137 /* Determine if an unsigned value is a power of two.
140 * Zero is \b not treated as a power of two.
143 util_is_power_of_two_nonzero(unsigned v
)
145 /* __POPCNT__ is different from HAVE___BUILTIN_POPCOUNT. The latter
146 * indicates the existence of the __builtin_popcount function. The former
147 * indicates that _mm_popcnt_u32 exists and is a native instruction.
149 * The other alternative is to use SSE 4.2 compile-time flags. This has
150 * two drawbacks. First, there is currently no build infrastructure for
151 * SSE 4.2 (only 4.1), so that would have to be added. Second, some AMD
152 * CPUs support POPCNT but not SSE 4.2 (e.g., Barcelona).
155 return _mm_popcnt_u32(v
) == 1;
157 return v
!= 0 && (v
& (v
- 1)) == 0;
161 /* For looping over a bitmask when you want to loop over consecutive bits
162 * manually, for example:
165 * int start, count, i;
167 * u_bit_scan_consecutive_range(&mask, &start, &count);
169 * for (i = 0; i < count; i++)
170 * ... process element (start+i)
174 u_bit_scan_consecutive_range(unsigned *mask
, int *start
, int *count
)
176 if (*mask
== 0xffffffff) {
182 *start
= ffs(*mask
) - 1;
183 *count
= ffs(~(*mask
>> *start
)) - 1;
184 *mask
&= ~(((1u << *count
) - 1) << *start
);
188 u_bit_scan_consecutive_range64(uint64_t *mask
, int *start
, int *count
)
190 if (*mask
== ~0ull) {
196 *start
= ffsll(*mask
) - 1;
197 *count
= ffsll(~(*mask
>> *start
)) - 1;
198 *mask
&= ~(((((uint64_t)1) << *count
) - 1) << *start
);
203 * Find last bit set in a word. The least significant bit is 1.
204 * Return 0 if no bits are set.
205 * Essentially ffs() in the reverse direction.
207 static inline unsigned
208 util_last_bit(unsigned u
)
210 #if defined(HAVE___BUILTIN_CLZ)
211 return u
== 0 ? 0 : 32 - __builtin_clz(u
);
212 #elif defined(_MSC_VER) && (_M_IX86 || _M_ARM || _M_AMD64 || _M_IA64)
214 if (_BitScanReverse(&index
, u
))
229 * Find last bit set in a word. The least significant bit is 1.
230 * Return 0 if no bits are set.
231 * Essentially ffsll() in the reverse direction.
233 static inline unsigned
234 util_last_bit64(uint64_t u
)
236 #if defined(HAVE___BUILTIN_CLZLL)
237 return u
== 0 ? 0 : 64 - __builtin_clzll(u
);
238 #elif defined(_MSC_VER) && (_M_AMD64 || _M_ARM || _M_IA64)
240 if (_BitScanReverse64(&index
, u
))
255 * Find last bit in a word that does not match the sign bit. The least
256 * significant bit is 1.
257 * Return 0 if no bits are set.
259 static inline unsigned
260 util_last_bit_signed(int i
)
263 return util_last_bit(i
);
265 return util_last_bit(~(unsigned)i
);
268 /* Returns a bitfield in which the first count bits starting at start are
271 static inline unsigned
272 u_bit_consecutive(unsigned start
, unsigned count
)
274 assert(start
+ count
<= 32);
277 return ((1u << count
) - 1) << start
;
280 static inline uint64_t
281 u_bit_consecutive64(unsigned start
, unsigned count
)
283 assert(start
+ count
<= 64);
286 return (((uint64_t)1 << count
) - 1) << start
;
294 #endif /* BITSCAN_H */