Computing the number of digits of an integer quickly

Suppose I give you an integer. How many decimal digits would you need to write it out? The number ‘100’ takes 3 digits whereas the number ’99’ requires only two.

You are effectively trying to compute the integer logarithm in base 10 of the number. I say ‘integer logarithm’ because you need to round up to the nearest integer.

Computers represent numbers in binary form, so it is easy to count the logarithm in base two. In C using GCC or clang, you can do so as follows using a counting leading zeroes function:

int int_log2(uint32_t x) { return 31 - __builtin_clz(x|1); }

Though it looks ugly, it is efficient. Most optimizing compilers on most systems will turn this into a single instruction.

How do you convert the logarithm in base 2 into the logarithm in base 10? From elementary mathematics, we have that log10 (x) = log2(x) / log2(10). So all you need is to divide by log2(10)… or get close enough. You do not want to actually divide, especially not by a floating-point value, so you want mutiply and shift instead. Multiplying and shifting is a standard technique to emulate the division.

You can get pretty close to a division by log2(10) if you multiply by 9 and then divide by 32 (2 to the power of 5). The division by a power of two is just a shift. (I initially used a division by a much larger power but readers corrected me.)

Unfortunately, that is not quite good enough because we do not actually have the logarithm in base 2, but rather a truncated version of it. Thus you may need to do a off-by-one correction. The following code works:

    static uint32_t table[] = {9, 99, 999, 9999, 99999, 
    999999, 9999999, 99999999, 999999999};
    int y = (9 * int_log2(x)) >> 5;
    y += x > table[y];
    return y + 1;

It might compile to the following assembly:

        or      eax, 1
        bsr     eax, eax
        lea     eax, [rax + 8*rax]
        shr     eax, 5
        cmp     dword ptr [4*rax + table], edi
        adc     eax, 0
        add     eax, 1

Loading from the table probably incurs multiple cycles of latency (e.g., 3 or 4). The x64 bsr instruction has also a long latency of 3 or 4 cycles. My code is available.

You can port this function to Java as follows if you assume that the number is non-negative:

        int l2 = 31 - Integer.numberOfLeadingZeros(num|1);
        int ans = ((9*l2)>>>5);
        if (num > table[ans]) { ans += 1; }
        return ans + 1;

I wrote this blog post to answer a question by Chris Seaton on Twitter. After writing it up, I found that the always-brilliant Travis Downs had proposed a similar solution with a table lookup. I believe he requires a larger table. Robert Clausecker once posted a solution that might be close to what Travis has in mind.

Furthermore, if the number of digits is predictable, then you can write code with branches and get better results in some cases. However, you should be concerned with the fact that a single branch miss can cost you 15 cycles and  tens of instructions.

Update: There is a followup to this blog post… Computing the number of digits of an integer even faster

Further reading: Converting binary integers to ASCII character and Integer log 10 of an unsigned integer — SIMD version

Note: Victor Zverovich stated on Twitter than the fmt C++ library relies on a similar approach. Pete Cawley showed that you could achieve the same result that I got initially by multiplying by 77 and then shifting by 8, instead of my initially larger constants. He implemented his solution for LuaJIt. Giulietti pointed out to me by email that almost exactly the same routine appears in Hacker’s Delight at the end of chapter 11.

Daniel Lemire, "Computing the number of digits of an integer quickly," in Daniel Lemire's blog, May 28, 2021.

Published by

Daniel Lemire

A computer science professor at the University of Quebec (TELUQ).

32 thoughts on “Computing the number of digits of an integer quickly”

  1. I have arrived at n=floor( log10(integer) +1 )
    Finding a proof of this formula would be interesting. I guess derivation of it is the proof itself.

    python version is, of course, len(str(integer))

    1. I have arrived at n=floor( log10(integer) +1 )

      Note that this cannot work for integers more than 2^53 or so, which cannot be represented exactly by IEEE double. The IEEE double-precision format can’t distinguish between 10^17 – 1 = ‭0x16345785d89ffff‬ and 10^17 = ‭0x16345785d8a0000‬. (In fact, it doesn’t work for me for 10^15 – 1, but that depends on the vagaries of the log10() implementation.)

      (Also, floating-point logarithms are not quick.)

  2. I found that the assembly of

    int digit_count(uint32_t x) {
    static uint32_t table[] = {9, 99, 999, 9999, 99999, 999999, 9999999, 99999999, 999999999};
    int y = (9 * (31 - __builtin_clz(x|1))) >> 5;
    y += x > table[y];
    return y + 1;

    looks a slightly faster, because the multiply by 9 can be done using lea instead of imul:

  3. Thanks for ruining my Saturday morning. 😛

    Here’s a digit count that takes a slightly different approach: Fix up the input to int_log2 so that the rounding doesn’t matter. When compiled with clang 12.0.0, it shaves an extra instruction off. Note that to get good compiled code, I had to add -O3 and change the definition of int_log2 a bit. (Two changes: Move |1 to the caller for more control, and use 31^ instead of 31- so that the compiler could do a better elimination job.)

    int int_log2(uint32_t x) { return 31 ^ __builtin_clz(x); }

    int digit_count(uint32_t x) {
    static uint32_t table[] = {
    0, 0, 0, (1<<4) - 10,
    0, 0, (1<<7) - 100,
    0, 0, (1<<10) - 1000,
    0, 0, 0, (1<<14) - 10000,
    0, 0, (1<<17) - 100000,
    0, 0, (1<<20) - 1000000,
    0, 0, 0, (1<<24) - 10000000,
    0, 0, (1<<27) - 100000000,
    0, 0, (1<<30) - 1000000000,
    0, 0,
    int l2 = int_log2(x);
    x += table[l2];
    l2 = int_log2(x|1);
    int ans = (77*l2)>>8;
    return ans + 1;

    Perhaps it can be improved further.

    I also have a reasonably optimized Go implementation that I’ll probably publish soon. (I’ll leave a note here if/when I do.)

    1. This us, unfortunately, silly. You’re calling ilog2 twice and using a larger table (one entry per bit rather than one per digit). All to save one conditional increment (e.g. compare, SETcc, add, or subtract, shift-right, add).

      1. It may be many things, but I do not believe it is silly.

        On my M1 laptop, it performs incrementally better than the post’s code. To run the entire 32 bit range, the 9*x>>5 approach takes 2.75s; this takes 2.67s.

        On my 2016 intel laptop, it performs worse, but I believe that that is because of a compiler shortcoming. When I use inline assembly thus:

        int in_asm(uint32_t x) {
        static uint64_t table[] = {16, 14, 12, 246, 240, 224, 3996, 3968, 3840, 64536, 64512, 63488, 61440, 1038576, 1032192, 1015808, 16677216, 16646144, 16515072, 267435456, 267386880, 266338304, 264241152, 4284967296, 4278190080, 4261412864, 68619476736, 68585259008, 68451041280, 1098511627776, 1098437885952, 1097364144128};
        int ret;
        __asm__ (
        "movl %%eax, %%edx;"
        "orl $1, %%edx;"
        "bsrl %%edx, %%edx;"
        "addq (%%rbx,%%rdx,8), %%rax;"
        "bsrq %%rax, %%rax;"
        "sarl $2, %%eax;"
        : "=a" (ret)
        : "a" (x), "b" (table)
        : "dx"
        return ret;

        I get the original code takes 4.09s to process all 32 bit integers, and the asm takes 3.03s. (This isn’t entirely fair, as I made no attempt to hand-optimize the original code. But I think it is enough to establish that this isn’t silly.)

        It is unfortunate that BSR has such a significant latency.

        Even if this wasn’t faster at all, I think it is interesting, because it is a fairly different kind of approach. It stemmed from the observation that the lookup table isn’t doing that much work. I was curious whether the lookup table could eliminate much of the extra calculations, which it turns out it can (no increments, lets us use x / 4 instead of 9x / 32).

        1. It looks like AMD processors (Zen+, Zen2 and Zen3) have the lzcnt instruction with a short (1 cycle) latency.

          I was surprised at the 3 cycle latency for bsr. I somehow assumed in my head that it was a 1-cycle instruction.

  4. Just for fun I would try this approach… use : log10(x) = log2(x) / log2(10)

    const float Recip_Log2_10 = 1f / log2( 10f );

    digit_count( x ) = 1 + (int)( log2(x) * Recip_Log2_10 )

    But instead of floating math do it in fixed-point?

    const int Recip_Log2_10 = (int)(( 1f / log2( 10f ) << 24 )

    digit_count( x ) = 1 + ( log2(x) * Recip_Log2_10 ) >> 24

    Its is on 99% total bulshit or at least 2 order slower… but that me 🙂

  5. Congrats, you ruined Sunday, too. 🙂

    I got it down to 5 instructions! (Well, it should be five, but clang and gcc each emit 6, because each sees an optimization the other missed.)

    bsr eax, edi
    mov ecx, edi
    add rcx, qword ptr [8*rax + j_digit_count.table]
    bsr rax, rcx
    sar eax, 2

    (gcc emits two movs after the first bsr; clang emits xor shr xor instead of sar.)

    Code is:

    int int_log2_64(uint64_t x) { return 63 ^ __builtin_clzll(x); }

    int digit_count(uint32_t x) {
    static uint64_t table[] = {16, 14, 12, 246, 240, 224, 3996, 3968, 3840, 64536, 64512, 63488, 61440, 1038576, 1032192, 1015808, 16677216, 16646144, 16515072, 267435456, 267386880, 266338304, 264241152, 4284967296, 4278190080, 4261412864, 68619476736, 68585259008, 68451041280, 1098511627776, 1098437885952, 1097364144128};
    int lg2 = int_log2(x);
    uint64_t n = (uint64_t)(x) + table[lg2];
    return int_log2_64(n) >> 2;

    I plan to write up an explanation of how this works on my own blog post soon(ish).

    1. mov ecx, edi is redundant there and you could probably get rid of it by casting to 32-bits in a strategic place. It’s just a quirk of the ABI that the high bits might contain garbage but in this case it’s harmless because it’s used as a shift count which has an implicit mask anyway.

      1. Nevermind, it’s not redundant, but it is just a quirk of the x86-64 ABI for passing 32-bit values in a 64-bit register. You could expect it to go away after inlining or if you made the parameter x a 64-bit value (but the top 32 bits must be zero).

    2. I plan to write up an explanation of how this works on my own blog
      post soon(ish).

      Please either ping me back or just add a comment here… so I can edit the blog post to refer to your blog post. (We want people who arrive at my blog post to find yours.)

    3. I see – you’re moving the step in decimal log to where the step in binary (actually base 16) log is. This technique should work for any set of ranges where log10 changes by at most 1 in each, ie where the end is less than 10 times the start.

      If we work with log base 8 for example, we can have one entry for 1-7, the next for 8-63, 64-511, and so on.

  6. Where the “if (v >= p10)” can result in an a slow branch, then (9 * l2) >>> 5 might not be best. The following results in fewer cases of “ans++” over the whole integer range (116 million instead of 1.6 billion).

    int l2 = 64 - Long.numberOfLeadingZeros(num | 1);
    int ans = (21845 * l2) >>> 16;
    long p10 = POW_10[ans];
    if (num >= p10) {
    return ans;

      1. Yes, I also hope a conditional move is possible here. For this case, I assume it’s more important to have multiply by 5 (or 9, 17,…). Only if a conditional move isn’t possible, reducing the probability of branches makes sense.

      2. Well if the probability of needing the fixup is low enough, then it would be better to use a branch instead and move this part off of the critial path (control dependency vs a data dependency).

        This would be provide a nice latency improvement at the cost of the occasional mispredict.

          1. No, I think it does either, just sort of mentioning this for completeness since it did come up in some other approaches I considered.

            I don’t think any approach which is purely a function of lzcnt(num) can achieve low correction probability because too much information has been lost in the truncation.

  7. Last night I realized that Josh’s approach can be modified to use a simpler function after the table mapping. More specifically, the table entry has the integer log10 in the upper 32 bits, and the lower 32 generate a carry when the appropriate threshold is reached, eg the entry around 100 is (3<<32) -100. We only need a >>32 to pull out the integer log after the add:

    mov eax, edi
    bsr rcx, rax
    add rax, qword ptr [8*rcx + digit_count(unsigned int)::table]
    shr rax, 32


    1. Here is the source — pardon my macro:

      int int_log2_64(uint64_t x) { return 63 ^ __builtin_clzll(x); }

      // this increments the upper 32 bits (log10(T) - 1) when >= T is added
      #define K(T) (((sizeof(#T)-1)<<32) - T)

      int digit_count(uint32_t x) {

      static uint64_t table[] = {
      K(10),K(10),K(10), // 64
      K(100),K(100),K(100), // 512
      K(1000),K(1000),K(1000), // 4096
      K(10000),K(10000),K(10000), // 32k
      K(100000),K(100000),K(100000), // 256k
      K(1000000),K(1000000),K(1000000), // 2048k
      K(10000000),K(10000000),K(10000000), // 16M
      K(100000000),K(100000000),K(100000000), // 128M
      K(1000000000),K(1000000000),K(1000000000), // 1024M
      K(1000000000),K(1000000000) // 4B

      int lg2 = int_log2_64(x);

      uint64_t n = (uint64_t)(x) + table[lg2];

      return n >> 32;

  8. I am wondering if a fast solution exists for uint64_t numbers.

    I am also interested, if there is fast solution to count the digits with over-estimation. What I mean is let say input is 12345, then the output can be 5,6,7 or even 20.

    This is very handy, if you have a digit, you want to convert it to string, but you need to know in advance how much memory bytes you need to allocate for the string. Think about 1000’s of strings.

    I thought I can calculate using octal numbers, however it not as easy as it may sound.

Leave a Reply

Your email address will not be published.

You may subscribe to this blog by email.