Odd-Major Trick
Pradu
Nov 14, 2007
Very nice trick! Should speedup Buzz a bit with my lousy single-parameter popcounts. :)
GerdIsenberg
Nov 14, 2007
Hi Pradu,
I accidently unmade your last changes, while adding an anchor.
Still not too familar with this wiki-stuff.
Can you please retore it again?
Sorry,
Gerd
Pradu
Nov 14, 2007
Ok, I'll have it done in a few minutes.
Future or recent x86-64 processors (AMD K10 - SSE4a, Intel Nehalem - SSE4.2) provide a 64-bit popcount instruction [2], available via C++ compiler intrinsic [3][4][5][6] or inline assembly [7]. Despite different Intrinsic prototypes (_mm_popcnt_u64 vs. popcnt64), Intel and AMD popcnt instructions are binary compatible, have same encoding (F3 [REX] 0F B8 /r), and both require bit 23 set in RCX of the CPUID function 0000_0001h.
Table of Contents
Recurrence Relation
The recursive recurrence relation of population counts can be transformed to iteration as well, but lacks an arithmetical sum-formula:However, it is helpful to initialize a lookup table, i.e. for bytes:
Empty or Single?
Often one has to deal with sparsely populated or even empty bitboards. To determine whether a bitboard is empty or a single populated power of two value, one may use simple boolean statements rather than a complete population count.Empty Bitboards
To test a bitboard is empty, one compares it with zero, or use the logical not operator:The inverse condition (not empty) is of course
Single Populated Bitboards
If the bitboard is not empty, we can extract the LS1B to look whether it is equal with the bitboard. Alternatively and faster, we can reset the LS1B to look whether the bitboard becomes empty.One can skip the leading x != 0 condition to test popcount <= 1:
Again the inverse relation tests, whether a bitboard has more than one bit set:
An alternative approach to determine single populated sets, aka power of two values is based on Inclusive LS1B separation divided by two equals the ones' decrement [9]:
Loop-Approaches
Too Slow
Brute force adding all 64-bitsOf course, this is a slow algorithm, which might be improved by testing x not empty rather than i < 64. But unrolled in parallel prefix manner it already reminds on this one.
Brian Kernighan's way
Consecutively reset LS1B in a loop body and counting loop cycles until the bitset becomes empty. Brian Kernighan [10] mentioned the trick in his and Ritchie's book The C Programming_Language, 2nd Edition 1988, exercise 2-9. However, the method was first published in 1960 by Peter Wegner [11], discovered independently by Derrick Henry Lehmer, published in 1964 [12]:Despite branches - this is still one of the fastest approaches for sparsely populated bitboards. Of course the more bits that are set, the longer it takes.
Lookup
Of course we can not use the whole bitboard as index to a lookup table - since it's size would be 18,446,744,073,709,551,616 bytes! If it is about the population count of a byte, we can simply construct a table lookup with 256 elements. For a bitboard that takes eight byte lookups we can add together:Looks quite expensive - one may use four 16-bit word-lookups with a pre-calculated 64KByte table though, but that pollutes the memory caches quite a bit. One can also treat the bitboard as array of bytes or words in memory, since endian issues don't care here - that safes all the shifts and 'ands', but has to read byte for byte from memory.
SWAR-Popcount
The divide and conquer SWAR-approach deals with counting bits of duos, to aggregate the duo-counts to nibbles and bytes inside one 64-bit register in parallel, to finally sum all bytes together. According to Donald Knuth [13], a parallel population count routine was already introduced in 1957 due to Donald B. Gillies and Jeffrey C. P. Miller in the first textbook on programming, second edition: The Preparation of Programs for an Electronic Digital Computer, by Maurice Wilkes, David Wheeler and Stanley Gill, pages 191–193 [14] [15].Counting Duo-Bits
A bit-duo (two neighboring bits) can be interpreted with bit 0 = a, and bit 1 = b asThe duo population is
which can be archived by
(2b+a)−(2b+a)÷2
orThe bit-duo has up to four states with population count from zero to two as demonstrated in following table with binary digits:
Note that the popcount-result of the bit-duos still takes two bits.
Counting Nibble-Bits
The next step is to add the duo-counts to populations of four neighboring bits, the 16 nibble-counts, which may range from zero to four. SWAR-wise it is done by masking odd and even duo-counts to add them together:Note that the popcount-result of the nibbles takes only three bits, since 100B is the maximum population (of the nibble 1111B).
Byte-Counts
You already got the idea? Now it is about to get the byte-populations from two nibble-populations. Maximum byte-population of 1000B only takes four bits, so it is safe to mask all those four bits of the sum, rather than to mask the summands:Adding the Byte-Counts
Parallel Prefix Adds
We may continue with mask-less parallel prefix SWAR-adds for byte-counts, word-counts and finally double-word-counts, to mask the least significant 8 (or 7) bits for final result in the 0..64 range:Multiplication
With todays fast 64-bit multiplication one can multiply the vector of 8-byte-counts with 0x0101010101010101 to get the final result in the most significant byte, which is then shifted right:Casting out
Interestingly, there is another approach to add the bytes together. As demonstrated with decimal digits (base 10) and Casting out nines [16], casting out by modulo base minus one is equivalent to taking the digit sum, which might be applied here with low range 0..8 "base 256" digits:However, since division and modulo are usually slow instructions and modulo by constant is likely replaced by reciprocal multiplication anyway by the compiler, the multiplication by 0x0101010101010101 aka the 2-adic fraction -1/255 is the preferred method.
The PopCount routine
The Constants
Putting all together, the various SWAR-Masks and factors as defined by Donald Knuth as 2-adic fractions [17]:represented as bitboards:
popCount
This is how the complete routine looks in C:Advantage: no branches, no memory lookups, constant runtime - independent from population
Drawback: dependency chain, not much parallel speedup
slowmul_popCount
And as stated before, for computers with relatively slower multiplication, the addition can be done manually:For likely sparsely populated bitboards, the loop-wise Brian Kernighan's way might be the faster one.
HAKMEM 169
A similar technique was proposed by Bill Gosper et al. from Massachusetts Institute of Technology, as published 1972 in Memo 239 (HAKMEM) [18] [19], to add bit-trio- rather than duo populations consecutively, and the 32 bit version relies on casting out 63. Note that the constants in the code below have octal (base-8) digits, originally written in Assembly for the PDP-6 [20]. An expanded 64-bit version, casting out 511 or 4095, is slightly less efficient than the binary SWAR version above.Miscellaneous
Cardinality of Multiple Sets
If we like to count arrays of sets, we can reduce 2^N-1 popcounts to N popcounts, by applying the odd-major-trick. For three sets to count we safe one, with five additional cheap instructions.The combined popCount3 likely gains more parallel speedup, since there are two independent chains to calculate. Possible Application is to pass the union of both bishops (since usually bishops cover disjoint sets due to different square colors) plus the up to two knight move-target sets.
Odd and Major 7-15
Of course - reducing seven popcount to three, or even 15 popcounts to four sounds even more promising.For N = 2^n - 1 it takes N - n odd-major pairs. Thus one add-major pair - five instructions - per saved popCount.
That is 7 - 3 = 4 pairs:
Or 15 - 4 = 11 pairs:
Odd and Major Digit Counts
Odd-Major is probably also useful to determine digit count sets of attacks or other stuff:with following semantics:
Popcount as log2 of LS1B
Assuming an architecture has a fast popcount-instruction (but no bitscan). One can isolate the LS1B, decrement it and count the remaining trailing "ones" to perform the logarithm dualis:For instance, LS1B is 2^44, decrementing leaves a below LSB1 mask with exactly 44 bits set:
Hamming Distance
The hamming distance of two words is defined as the number of corresponding different bits.Hamming distance greater than one or two is an important property of codes to detect or even correct one-bit errors.
The minimum and average hamming distance over all Zobrist keys was considered as "quality"-measure of the keys. However, as long the minimum hamming distance is greater zero, linear independence (that is a small subset of all keys doesn't xor to zero), is much more important than hamming distance [21]. Maximizing the minimal hamming distance leads to very poor Zobrist keys [22].
Weighted PopCount
For a SIMD-wise kind of weighted population count, see the SSE2 dot-product.Pre-calculated Mobility
Similar to Attacks by Occupancy Lookup to determine attack sets of sliding pieces, we may use pre-calculated population count or even center-weighted population count as a rough estimate on piece mobility [23]. It may not consider subsets of let say safe target squares.Piece Attacks Count
As pointed out by Marco Costalba [24] [25], specialized routines to count the population (Mobility) of attack sets of king, knight and line-wise sub-sets of sliding pieces can be done more efficiently than the general SWAR-Popcount. This is similar to Flipping Mirroring and Rotating the whole bitboard versus Rank, File and Diagonal, and is based on mapping the up to eight scattered occupied bits to one byte, to perform a single byte lookup. For various mapping techniques, see:Popcount in Hardware
See also
Publications
1949 ...
2000 ...
2010 ...
Postings
1998 ...
2000 ...
2005 ...
2010 ...
2015 ...
External Links
HAKMEMC -- HAKMEM Programming hacks in C by Alan Mycroft
John Abercrombie, Bobo Stenson, Lars Danielsson, Jon Christensen
References
What links here?
Up one Level