mirror of
				https://github.com/python/cpython.git
				synced 2025-11-04 11:49:12 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			729 lines
		
	
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			729 lines
		
	
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
"""Python implementations of some algorithms for use by longobject.c.
 | 
						|
The goal is to provide asymptotically faster algorithms that can be
 | 
						|
used for operations on integers with many digits.  In those cases, the
 | 
						|
performance overhead of the Python implementation is not significant
 | 
						|
since the asymptotic behavior is what dominates runtime. Functions
 | 
						|
provided by this module should be considered private and not part of any
 | 
						|
public API.
 | 
						|
 | 
						|
Note: for ease of maintainability, please prefer clear code and avoid
 | 
						|
"micro-optimizations".  This module will only be imported and used for
 | 
						|
integers with a huge number of digits.  Saving a few microseconds with
 | 
						|
tricky or non-obvious code is not worth it.  For people looking for
 | 
						|
maximum performance, they should use something like gmpy2."""
 | 
						|
 | 
						|
import re
 | 
						|
import decimal
 | 
						|
try:
 | 
						|
    import _decimal
 | 
						|
except ImportError:
 | 
						|
    _decimal = None
 | 
						|
 | 
						|
# A number of functions have this form, where `w` is a desired number of
 | 
						|
# digits in base `base`:
 | 
						|
#
 | 
						|
#    def inner(...w...):
 | 
						|
#        if w <= LIMIT:
 | 
						|
#            return something
 | 
						|
#        lo = w >> 1
 | 
						|
#        hi = w - lo
 | 
						|
#        something involving base**lo, inner(...lo...), j, and inner(...hi...)
 | 
						|
#    figure out largest w needed
 | 
						|
#    result = inner(w)
 | 
						|
#
 | 
						|
# They all had some on-the-fly scheme to cache `base**lo` results for reuse.
 | 
						|
# Power is costly.
 | 
						|
#
 | 
						|
# This routine aims to compute all amd only the needed powers in advance, as
 | 
						|
# efficiently as reasonably possible. This isn't trivial, and all the
 | 
						|
# on-the-fly methods did needless work in many cases. The driving code above
 | 
						|
# changes to:
 | 
						|
#
 | 
						|
#    figure out largest w needed
 | 
						|
#    mycache = compute_powers(w, base, LIMIT)
 | 
						|
#    result = inner(w)
 | 
						|
#
 | 
						|
# and `mycache[lo]` replaces `base**lo` in the inner function.
 | 
						|
#
 | 
						|
# If an algorithm wants the powers of ceiling(w/2) instead of the floor,
 | 
						|
# pass keyword argument `need_hi=True`.
 | 
						|
#
 | 
						|
# While this does give minor speedups (a few percent at best), the
 | 
						|
# primary intent is to simplify the functions using this, by eliminating
 | 
						|
# the need for them to craft their own ad-hoc caching schemes.
 | 
						|
#
 | 
						|
# See code near end of file for a block of code that can be enabled to
 | 
						|
# run millions of tests.
 | 
						|
def compute_powers(w, base, more_than, *, need_hi=False, show=False):
 | 
						|
    seen = set()
 | 
						|
    need = set()
 | 
						|
    ws = {w}
 | 
						|
    while ws:
 | 
						|
        w = ws.pop() # any element is fine to use next
 | 
						|
        if w in seen or w <= more_than:
 | 
						|
            continue
 | 
						|
        seen.add(w)
 | 
						|
        lo = w >> 1
 | 
						|
        hi = w - lo
 | 
						|
        # only _need_ one here; the other may, or may not, be needed
 | 
						|
        which = hi if need_hi else lo
 | 
						|
        need.add(which)
 | 
						|
        ws.add(which)
 | 
						|
        if lo != hi:
 | 
						|
            ws.add(w - which)
 | 
						|
 | 
						|
    # `need` is the set of exponents needed. To compute them all
 | 
						|
    # efficiently, possibly add other exponents to `extra`. The goal is
 | 
						|
    # to ensure that each exponent can be gotten from a smaller one via
 | 
						|
    # multiplying by the base, squaring it, or squaring and then
 | 
						|
    # multiplying by the base.
 | 
						|
    #
 | 
						|
    # If need_hi is False, this is already the case (w can always be
 | 
						|
    # gotten from w >> 1 via one of the squaring strategies). But we do
 | 
						|
    # the work anyway, just in case ;-)
 | 
						|
    #
 | 
						|
    # Note that speed is irrelevant. These loops are working on little
 | 
						|
    # ints (exponents) and go around O(log w) times. The total cost is
 | 
						|
    # insignificant compared to just one of the bigint multiplies.
 | 
						|
    cands = need.copy()
 | 
						|
    extra = set()
 | 
						|
    while cands:
 | 
						|
        w = max(cands)
 | 
						|
        cands.remove(w)
 | 
						|
        lo = w >> 1
 | 
						|
        if lo > more_than and w-1 not in cands and lo not in cands:
 | 
						|
            extra.add(lo)
 | 
						|
            cands.add(lo)
 | 
						|
    assert need_hi or not extra
 | 
						|
 | 
						|
    d = {}
 | 
						|
    for n in sorted(need | extra):
 | 
						|
        lo = n >> 1
 | 
						|
        hi = n - lo
 | 
						|
        if n-1 in d:
 | 
						|
            if show:
 | 
						|
                print("* base", end="")
 | 
						|
            result = d[n-1] * base # cheap!
 | 
						|
        elif lo in d:
 | 
						|
            # Multiplying a bigint by itself is about twice as fast
 | 
						|
            # in CPython provided it's the same object.
 | 
						|
            if show:
 | 
						|
                print("square", end="")
 | 
						|
            result = d[lo] * d[lo] # same object
 | 
						|
            if hi != lo:
 | 
						|
                if show:
 | 
						|
                    print(" * base", end="")
 | 
						|
                assert 2 * lo + 1 == n
 | 
						|
                result *= base
 | 
						|
        else: # rare
 | 
						|
            if show:
 | 
						|
                print("pow", end='')
 | 
						|
            result = base ** n
 | 
						|
        if show:
 | 
						|
            print(" at", n, "needed" if n in need else "extra")
 | 
						|
        d[n] = result
 | 
						|
 | 
						|
    assert need <= d.keys()
 | 
						|
    if excess := d.keys() - need:
 | 
						|
        assert need_hi
 | 
						|
        for n in excess:
 | 
						|
            del d[n]
 | 
						|
    return d
 | 
						|
 | 
						|
_unbounded_dec_context = decimal.getcontext().copy()
 | 
						|
_unbounded_dec_context.prec = decimal.MAX_PREC
 | 
						|
_unbounded_dec_context.Emax = decimal.MAX_EMAX
 | 
						|
_unbounded_dec_context.Emin = decimal.MIN_EMIN
 | 
						|
_unbounded_dec_context.traps[decimal.Inexact] = 1 # sanity check
 | 
						|
 | 
						|
def int_to_decimal(n):
 | 
						|
    """Asymptotically fast conversion of an 'int' to Decimal."""
 | 
						|
 | 
						|
    # Function due to Tim Peters.  See GH issue #90716 for details.
 | 
						|
    # https://github.com/python/cpython/issues/90716
 | 
						|
    #
 | 
						|
    # The implementation in longobject.c of base conversion algorithms
 | 
						|
    # between power-of-2 and non-power-of-2 bases are quadratic time.
 | 
						|
    # This function implements a divide-and-conquer algorithm that is
 | 
						|
    # faster for large numbers.  Builds an equal decimal.Decimal in a
 | 
						|
    # "clever" recursive way.  If we want a string representation, we
 | 
						|
    # apply str to _that_.
 | 
						|
 | 
						|
    from decimal import Decimal as D
 | 
						|
    BITLIM = 200
 | 
						|
 | 
						|
    # Don't bother caching the "lo" mask in this; the time to compute it is
 | 
						|
    # tiny compared to the multiply.
 | 
						|
    def inner(n, w):
 | 
						|
        if w <= BITLIM:
 | 
						|
            return D(n)
 | 
						|
        w2 = w >> 1
 | 
						|
        hi = n >> w2
 | 
						|
        lo = n & ((1 << w2) - 1)
 | 
						|
        return inner(lo, w2) + inner(hi, w - w2) * w2pow[w2]
 | 
						|
 | 
						|
    with decimal.localcontext(_unbounded_dec_context):
 | 
						|
        nbits = n.bit_length()
 | 
						|
        w2pow = compute_powers(nbits, D(2), BITLIM)
 | 
						|
        if n < 0:
 | 
						|
            negate = True
 | 
						|
            n = -n
 | 
						|
        else:
 | 
						|
            negate = False
 | 
						|
        result = inner(n, nbits)
 | 
						|
        if negate:
 | 
						|
            result = -result
 | 
						|
    return result
 | 
						|
 | 
						|
def int_to_decimal_string(n):
 | 
						|
    """Asymptotically fast conversion of an 'int' to a decimal string."""
 | 
						|
    w = n.bit_length()
 | 
						|
    if w > 450_000 and _decimal is not None:
 | 
						|
        # It is only usable with the C decimal implementation.
 | 
						|
        # _pydecimal.py calls str() on very large integers, which in its
 | 
						|
        # turn calls int_to_decimal_string(), causing very deep recursion.
 | 
						|
        return str(int_to_decimal(n))
 | 
						|
 | 
						|
    # Fallback algorithm for the case when the C decimal module isn't
 | 
						|
    # available.  This algorithm is asymptotically worse than the algorithm
 | 
						|
    # using the decimal module, but better than the quadratic time
 | 
						|
    # implementation in longobject.c.
 | 
						|
 | 
						|
    DIGLIM = 1000
 | 
						|
    def inner(n, w):
 | 
						|
        if w <= DIGLIM:
 | 
						|
            return str(n)
 | 
						|
        w2 = w >> 1
 | 
						|
        hi, lo = divmod(n, pow10[w2])
 | 
						|
        return inner(hi, w - w2) + inner(lo, w2).zfill(w2)
 | 
						|
 | 
						|
    # The estimation of the number of decimal digits.
 | 
						|
    # There is no harm in small error.  If we guess too large, there may
 | 
						|
    # be leading 0's that need to be stripped.  If we guess too small, we
 | 
						|
    # may need to call str() recursively for the remaining highest digits,
 | 
						|
    # which can still potentially be a large integer. This is manifested
 | 
						|
    # only if the number has way more than 10**15 digits, that exceeds
 | 
						|
    # the 52-bit physical address limit in both Intel64 and AMD64.
 | 
						|
    w = int(w * 0.3010299956639812 + 1)  # log10(2)
 | 
						|
    pow10 = compute_powers(w, 5, DIGLIM)
 | 
						|
    for k, v in pow10.items():
 | 
						|
        pow10[k] = v << k # 5**k << k == 5**k * 2**k == 10**k
 | 
						|
    if n < 0:
 | 
						|
        n = -n
 | 
						|
        sign = '-'
 | 
						|
    else:
 | 
						|
        sign = ''
 | 
						|
    s = inner(n, w)
 | 
						|
    if s[0] == '0' and n:
 | 
						|
        # If our guess of w is too large, there may be leading 0's that
 | 
						|
        # need to be stripped.
 | 
						|
        s = s.lstrip('0')
 | 
						|
    return sign + s
 | 
						|
 | 
						|
def _str_to_int_inner(s):
 | 
						|
    """Asymptotically fast conversion of a 'str' to an 'int'."""
 | 
						|
 | 
						|
    # Function due to Bjorn Martinsson.  See GH issue #90716 for details.
 | 
						|
    # https://github.com/python/cpython/issues/90716
 | 
						|
    #
 | 
						|
    # The implementation in longobject.c of base conversion algorithms
 | 
						|
    # between power-of-2 and non-power-of-2 bases are quadratic time.
 | 
						|
    # This function implements a divide-and-conquer algorithm making use
 | 
						|
    # of Python's built in big int multiplication. Since Python uses the
 | 
						|
    # Karatsuba algorithm for multiplication, the time complexity
 | 
						|
    # of this function is O(len(s)**1.58).
 | 
						|
 | 
						|
    DIGLIM = 2048
 | 
						|
 | 
						|
    def inner(a, b):
 | 
						|
        if b - a <= DIGLIM:
 | 
						|
            return int(s[a:b])
 | 
						|
        mid = (a + b + 1) >> 1
 | 
						|
        return (inner(mid, b)
 | 
						|
                + ((inner(a, mid) * w5pow[b - mid])
 | 
						|
                    << (b - mid)))
 | 
						|
 | 
						|
    w5pow = compute_powers(len(s), 5, DIGLIM)
 | 
						|
    return inner(0, len(s))
 | 
						|
 | 
						|
 | 
						|
# Asymptotically faster version, using the C decimal module. See
 | 
						|
# comments at the end of the file. This uses decimal arithmetic to
 | 
						|
# convert from base 10 to base 256. The latter is just a string of
 | 
						|
# bytes, which CPython can convert very efficiently to a Python int.
 | 
						|
 | 
						|
# log of 10 to base 256 with best-possible 53-bit precision. Obtained
 | 
						|
# via:
 | 
						|
#    from mpmath import mp
 | 
						|
#    mp.prec = 1000
 | 
						|
#    print(float(mp.log(10, 256)).hex())
 | 
						|
_LOG_10_BASE_256 = float.fromhex('0x1.a934f0979a371p-2') # about 0.415
 | 
						|
 | 
						|
# _spread is for internal testing. It maps a key to the number of times
 | 
						|
# that condition obtained in _dec_str_to_int_inner:
 | 
						|
#     key 0 - quotient guess was right
 | 
						|
#     key 1 - quotient had to be boosted by 1, one time
 | 
						|
#     key 999 - one adjustment wasn't enough, so fell back to divmod
 | 
						|
from collections import defaultdict
 | 
						|
_spread = defaultdict(int)
 | 
						|
del defaultdict
 | 
						|
 | 
						|
def _dec_str_to_int_inner(s, *, GUARD=8):
 | 
						|
    # Yes, BYTELIM is "large". Large enough that CPython will usually
 | 
						|
    # use the Karatsuba _str_to_int_inner to convert the string. This
 | 
						|
    # allowed reducing the cutoff for calling _this_ function from 3.5M
 | 
						|
    # to 2M digits. We could almost certainly do even better by
 | 
						|
    # fine-tuning this and/or using a larger output base than 256.
 | 
						|
    BYTELIM = 100_000
 | 
						|
    D = decimal.Decimal
 | 
						|
    result = bytearray()
 | 
						|
    # See notes at end of file for discussion of GUARD.
 | 
						|
    assert GUARD > 0 # if 0, `decimal` can blow up - .prec 0 not allowed
 | 
						|
 | 
						|
    def inner(n, w):
 | 
						|
        #assert n < D256 ** w # required, but too expensive to check
 | 
						|
        if w <= BYTELIM:
 | 
						|
            # XXX Stefan Pochmann discovered that, for 1024-bit ints,
 | 
						|
            # `int(Decimal)` took 2.5x longer than `int(str(Decimal))`.
 | 
						|
            # Worse, `int(Decimal) is still quadratic-time for much
 | 
						|
            # larger ints. So unless/until all that is repaired, the
 | 
						|
            # seemingly redundant `str(Decimal)` is crucial to speed.
 | 
						|
            result.extend(int(str(n)).to_bytes(w)) # big-endian default
 | 
						|
            return
 | 
						|
        w1 = w >> 1
 | 
						|
        w2 = w - w1
 | 
						|
        if 0:
 | 
						|
            # This is maximally clear, but "too slow". `decimal`
 | 
						|
            # division is asymptotically fast, but we have no way to
 | 
						|
            # tell it to reuse the high-precision reciprocal it computes
 | 
						|
            # for pow256[w2], so it has to recompute it over & over &
 | 
						|
            # over again :-(
 | 
						|
            hi, lo = divmod(n, pow256[w2][0])
 | 
						|
        else:
 | 
						|
            p256, recip = pow256[w2]
 | 
						|
            # The integer part will have a number of digits about equal
 | 
						|
            # to the difference between the log10s of `n` and `pow256`
 | 
						|
            # (which, since these are integers, is roughly approximated
 | 
						|
            # by `.adjusted()`). That's the working precision we need,
 | 
						|
            ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD
 | 
						|
            hi = +n * +recip # unary `+` chops back to ctx.prec digits
 | 
						|
            ctx.prec = decimal.MAX_PREC
 | 
						|
            hi = hi.to_integral_value() # lose the fractional digits
 | 
						|
            lo = n - hi * p256
 | 
						|
            # Because we've been uniformly rounding down, `hi` is a
 | 
						|
            # lower bound on the correct quotient.
 | 
						|
            assert lo >= 0
 | 
						|
            # Adjust quotient up if needed. It usually isn't. In random
 | 
						|
            # testing on inputs through 5 billion digit strings, the
 | 
						|
            # test triggered once in about 200 thousand tries.
 | 
						|
            count = 0
 | 
						|
            if lo >= p256:
 | 
						|
                count = 1
 | 
						|
                lo -= p256
 | 
						|
                hi += 1
 | 
						|
                if lo >= p256:
 | 
						|
                    # Complete correction via an exact computation. I
 | 
						|
                    # believe it's not possible to get here provided
 | 
						|
                    # GUARD >= 3. It's tested by reducing GUARD below
 | 
						|
                    # that.
 | 
						|
                    count = 999
 | 
						|
                    hi2, lo = divmod(lo, p256)
 | 
						|
                    hi += hi2
 | 
						|
            _spread[count] += 1
 | 
						|
            # The assert should always succeed, but way too slow to keep
 | 
						|
            # enabled.
 | 
						|
            #assert hi, lo == divmod(n, pow256[w2][0])
 | 
						|
        inner(hi, w1)
 | 
						|
        del hi # at top levels, can free a lot of RAM "early"
 | 
						|
        inner(lo, w2)
 | 
						|
 | 
						|
    # How many base 256 digits are needed?. Mathematically, exactly
 | 
						|
    # floor(log256(int(s))) + 1. There is no cheap way to compute this.
 | 
						|
    # But we can get an upper bound, and that's necessary for our error
 | 
						|
    # analysis to make sense. int(s) < 10**len(s), so the log needed is
 | 
						|
    # < log256(10**len(s)) = len(s) * log256(10). However, using
 | 
						|
    # finite-precision floating point for this, it's possible that the
 | 
						|
    # computed value is a little less than the true value. If the true
 | 
						|
    # value is at - or a little higher than - an integer, we can get an
 | 
						|
    # off-by-1 error too low. So we add 2 instead of 1 if chopping lost
 | 
						|
    # a fraction > 0.9.
 | 
						|
 | 
						|
    # The "WASI" test platform can complain about `len(s)` if it's too
 | 
						|
    # large to fit in its idea of "an index-sized integer".
 | 
						|
    lenS = s.__len__()
 | 
						|
    log_ub = lenS * _LOG_10_BASE_256
 | 
						|
    log_ub_as_int = int(log_ub)
 | 
						|
    w = log_ub_as_int + 1 + (log_ub - log_ub_as_int > 0.9)
 | 
						|
    # And what if we've plain exhausted the limits of HW floats? We
 | 
						|
    # could compute the log to any desired precision using `decimal`,
 | 
						|
    # but it's not plausible that anyone will pass a string requiring
 | 
						|
    # trillions of bytes (unless they're just trying to "break things").
 | 
						|
    if w.bit_length() >= 46:
 | 
						|
        # "Only" had < 53 - 46 = 7 bits to spare in IEEE-754 double.
 | 
						|
        raise ValueError(f"cannot convert string of len {lenS} to int")
 | 
						|
    with decimal.localcontext(_unbounded_dec_context) as ctx:
 | 
						|
        D256 = D(256)
 | 
						|
        pow256 = compute_powers(w, D256, BYTELIM, need_hi=True)
 | 
						|
        rpow256 = compute_powers(w, 1 / D256, BYTELIM, need_hi=True)
 | 
						|
        # We're going to do inexact, chopped arithmetic, multiplying by
 | 
						|
        # an approximation to the reciprocal of 256**i. We chop to get a
 | 
						|
        # lower bound on the true integer quotient. Our approximation is
 | 
						|
        # a lower bound, the multiplication is chopped too, and
 | 
						|
        # to_integral_value() is also chopped.
 | 
						|
        ctx.traps[decimal.Inexact] = 0
 | 
						|
        ctx.rounding = decimal.ROUND_DOWN
 | 
						|
        for k, v in pow256.items():
 | 
						|
            # No need to save much more precision in the reciprocal than
 | 
						|
            # the power of 256 has, plus some guard digits to absorb
 | 
						|
            # most relevant rounding errors. This is highly significant:
 | 
						|
            # 1/2**i has the same number of significant decimal digits
 | 
						|
            # as 5**i, generally over twice the number in 2**i,
 | 
						|
            ctx.prec = v.adjusted() + GUARD + 1
 | 
						|
            # The unary "+" chops the reciprocal back to that precision.
 | 
						|
            pow256[k] = v, +rpow256[k]
 | 
						|
        del rpow256 # exact reciprocals no longer needed
 | 
						|
        ctx.prec = decimal.MAX_PREC
 | 
						|
        inner(D(s), w)
 | 
						|
    return int.from_bytes(result)
 | 
						|
 | 
						|
def int_from_string(s):
 | 
						|
    """Asymptotically fast version of PyLong_FromString(), conversion
 | 
						|
    of a string of decimal digits into an 'int'."""
 | 
						|
    # PyLong_FromString() has already removed leading +/-, checked for invalid
 | 
						|
    # use of underscore characters, checked that string consists of only digits
 | 
						|
    # and underscores, and stripped leading whitespace.  The input can still
 | 
						|
    # contain underscores and have trailing whitespace.
 | 
						|
    s = s.rstrip().replace('_', '')
 | 
						|
    func = _str_to_int_inner
 | 
						|
    if len(s) >= 2_000_000 and _decimal is not None:
 | 
						|
        func = _dec_str_to_int_inner
 | 
						|
    return func(s)
 | 
						|
 | 
						|
def str_to_int(s):
 | 
						|
    """Asymptotically fast version of decimal string to 'int' conversion."""
 | 
						|
    # FIXME: this doesn't support the full syntax that int() supports.
 | 
						|
    m = re.match(r'\s*([+-]?)([0-9_]+)\s*', s)
 | 
						|
    if not m:
 | 
						|
        raise ValueError('invalid literal for int() with base 10')
 | 
						|
    v = int_from_string(m.group(2))
 | 
						|
    if m.group(1) == '-':
 | 
						|
        v = -v
 | 
						|
    return v
 | 
						|
 | 
						|
 | 
						|
# Fast integer division, based on code from Mark Dickinson, fast_div.py
 | 
						|
# GH-47701. Additional refinements and optimizations by Bjorn Martinsson.  The
 | 
						|
# algorithm is due to Burnikel and Ziegler, in their paper "Fast Recursive
 | 
						|
# Division".
 | 
						|
 | 
						|
_DIV_LIMIT = 4000
 | 
						|
 | 
						|
 | 
						|
def _div2n1n(a, b, n):
 | 
						|
    """Divide a 2n-bit nonnegative integer a by an n-bit positive integer
 | 
						|
    b, using a recursive divide-and-conquer algorithm.
 | 
						|
 | 
						|
    Inputs:
 | 
						|
      n is a positive integer
 | 
						|
      b is a positive integer with exactly n bits
 | 
						|
      a is a nonnegative integer such that a < 2**n * b
 | 
						|
 | 
						|
    Output:
 | 
						|
      (q, r) such that a = b*q+r and 0 <= r < b.
 | 
						|
 | 
						|
    """
 | 
						|
    if a.bit_length() - n <= _DIV_LIMIT:
 | 
						|
        return divmod(a, b)
 | 
						|
    pad = n & 1
 | 
						|
    if pad:
 | 
						|
        a <<= 1
 | 
						|
        b <<= 1
 | 
						|
        n += 1
 | 
						|
    half_n = n >> 1
 | 
						|
    mask = (1 << half_n) - 1
 | 
						|
    b1, b2 = b >> half_n, b & mask
 | 
						|
    q1, r = _div3n2n(a >> n, (a >> half_n) & mask, b, b1, b2, half_n)
 | 
						|
    q2, r = _div3n2n(r, a & mask, b, b1, b2, half_n)
 | 
						|
    if pad:
 | 
						|
        r >>= 1
 | 
						|
    return q1 << half_n | q2, r
 | 
						|
 | 
						|
 | 
						|
def _div3n2n(a12, a3, b, b1, b2, n):
 | 
						|
    """Helper function for _div2n1n; not intended to be called directly."""
 | 
						|
    if a12 >> n == b1:
 | 
						|
        q, r = (1 << n) - 1, a12 - (b1 << n) + b1
 | 
						|
    else:
 | 
						|
        q, r = _div2n1n(a12, b1, n)
 | 
						|
    r = (r << n | a3) - q * b2
 | 
						|
    while r < 0:
 | 
						|
        q -= 1
 | 
						|
        r += b
 | 
						|
    return q, r
 | 
						|
 | 
						|
 | 
						|
def _int2digits(a, n):
 | 
						|
    """Decompose non-negative int a into base 2**n
 | 
						|
 | 
						|
    Input:
 | 
						|
      a is a non-negative integer
 | 
						|
 | 
						|
    Output:
 | 
						|
      List of the digits of a in base 2**n in little-endian order,
 | 
						|
      meaning the most significant digit is last. The most
 | 
						|
      significant digit is guaranteed to be non-zero.
 | 
						|
      If a is 0 then the output is an empty list.
 | 
						|
 | 
						|
    """
 | 
						|
    a_digits = [0] * ((a.bit_length() + n - 1) // n)
 | 
						|
 | 
						|
    def inner(x, L, R):
 | 
						|
        if L + 1 == R:
 | 
						|
            a_digits[L] = x
 | 
						|
            return
 | 
						|
        mid = (L + R) >> 1
 | 
						|
        shift = (mid - L) * n
 | 
						|
        upper = x >> shift
 | 
						|
        lower = x ^ (upper << shift)
 | 
						|
        inner(lower, L, mid)
 | 
						|
        inner(upper, mid, R)
 | 
						|
 | 
						|
    if a:
 | 
						|
        inner(a, 0, len(a_digits))
 | 
						|
    return a_digits
 | 
						|
 | 
						|
 | 
						|
def _digits2int(digits, n):
 | 
						|
    """Combine base-2**n digits into an int. This function is the
 | 
						|
    inverse of `_int2digits`. For more details, see _int2digits.
 | 
						|
    """
 | 
						|
 | 
						|
    def inner(L, R):
 | 
						|
        if L + 1 == R:
 | 
						|
            return digits[L]
 | 
						|
        mid = (L + R) >> 1
 | 
						|
        shift = (mid - L) * n
 | 
						|
        return (inner(mid, R) << shift) + inner(L, mid)
 | 
						|
 | 
						|
    return inner(0, len(digits)) if digits else 0
 | 
						|
 | 
						|
 | 
						|
def _divmod_pos(a, b):
 | 
						|
    """Divide a non-negative integer a by a positive integer b, giving
 | 
						|
    quotient and remainder."""
 | 
						|
    # Use grade-school algorithm in base 2**n, n = nbits(b)
 | 
						|
    n = b.bit_length()
 | 
						|
    a_digits = _int2digits(a, n)
 | 
						|
 | 
						|
    r = 0
 | 
						|
    q_digits = []
 | 
						|
    for a_digit in reversed(a_digits):
 | 
						|
        q_digit, r = _div2n1n((r << n) + a_digit, b, n)
 | 
						|
        q_digits.append(q_digit)
 | 
						|
    q_digits.reverse()
 | 
						|
    q = _digits2int(q_digits, n)
 | 
						|
    return q, r
 | 
						|
 | 
						|
 | 
						|
def int_divmod(a, b):
 | 
						|
    """Asymptotically fast replacement for divmod, for 'int'.
 | 
						|
    Its time complexity is O(n**1.58), where n = #bits(a) + #bits(b).
 | 
						|
    """
 | 
						|
    if b == 0:
 | 
						|
        raise ZeroDivisionError('division by zero')
 | 
						|
    elif b < 0:
 | 
						|
        q, r = int_divmod(-a, -b)
 | 
						|
        return q, -r
 | 
						|
    elif a < 0:
 | 
						|
        q, r = int_divmod(~a, b)
 | 
						|
        return ~q, b + ~r
 | 
						|
    else:
 | 
						|
        return _divmod_pos(a, b)
 | 
						|
 | 
						|
 | 
						|
# Notes on _dec_str_to_int_inner:
 | 
						|
#
 | 
						|
# Stefan Pochmann worked up a str->int function that used the decimal
 | 
						|
# module to, in effect, convert from base 10 to base 256. This is
 | 
						|
# "unnatural", in that it requires multiplying and dividing by large
 | 
						|
# powers of 2, which `decimal` isn't naturally suited to. But
 | 
						|
# `decimal`'s `*` and `/` are asymptotically superior to CPython's, so
 | 
						|
# at _some_ point it could be expected to win.
 | 
						|
#
 | 
						|
# Alas, the crossover point was too high to be of much real interest. I
 | 
						|
# (Tim) then worked on ways to replace its division with multiplication
 | 
						|
# by a cached reciprocal approximation instead, fixing up errors
 | 
						|
# afterwards. This reduced the crossover point significantly,
 | 
						|
#
 | 
						|
# I revisited the code, and found ways to improve and simplify it. The
 | 
						|
# crossover point is at about 3.4 million digits now.
 | 
						|
#
 | 
						|
# About .adjusted()
 | 
						|
# -----------------
 | 
						|
# Restrict to Decimal values x > 0. We don't use negative numbers in the
 | 
						|
# code, and I don't want to have to keep typing, e.g., "absolute value".
 | 
						|
#
 | 
						|
# For convenience, I'll use `x.a` to mean `x.adjusted()`. x.a doesn't
 | 
						|
# look at the digits of x, but instead returns an integer giving x's
 | 
						|
# order of magnitude. These are equivalent:
 | 
						|
#
 | 
						|
# - x.a is the power-of-10 exponent of x's most significant digit.
 | 
						|
# - x.a = the infinitely precise floor(log10(x))
 | 
						|
# - x can be written in this form, where f is a real with 1 <= f < 10:
 | 
						|
#    x = f * 10**x.a
 | 
						|
#
 | 
						|
# Observation; if x is an integer, len(str(x)) = x.a + 1.
 | 
						|
#
 | 
						|
# Lemma 1: (x * y).a = x.a + y.a, or one larger
 | 
						|
#
 | 
						|
# Proof: Write x = f * 10**x.a and y = g * 10**y.a, where f and g are in
 | 
						|
# [1, 10). Then x*y = f*g * 10**(x.a + y.a), where 1 <= f*g < 100. If
 | 
						|
# f*g < 10, (x*y).a is x.a+y.a. Else divide f*g by 10 to bring it back
 | 
						|
# into [1, 10], and add 1 to the exponent to compensate. Then (x*y).a is
 | 
						|
# x.a+y.a+1.
 | 
						|
#
 | 
						|
# Lemma 2: ceiling(log10(x/y)) <= x.a - y.a + 1
 | 
						|
#
 | 
						|
# Proof: Express x and y as in Lemma 1. Then x/y = f/g * 10**(x.a -
 | 
						|
# y.a), where 1/10 < f/g < 10. If 1 <= f/g, (x/y).a is x.a-y.a. Else
 | 
						|
# multiply f/g by 10 to bring it back into [1, 10], and subtract 1 from
 | 
						|
# the exponent to compensate. Then (x/y).a is x.a-y.a-1. So the largest
 | 
						|
# (x/y).a can be is x.a-y.a. Since that's the floor of log10(x/y). the
 | 
						|
# ceiling is at most 1 larger (with equality iff f/g = 1 exactly).
 | 
						|
#
 | 
						|
# GUARD digits
 | 
						|
# ------------
 | 
						|
# We only want the integer part of divisions, so don't need to build
 | 
						|
# the full multiplication tree. But using _just_ the number of
 | 
						|
# digits expected in the integer part ignores too much. What's left
 | 
						|
# out can have a very significant effect on the quotient. So we use
 | 
						|
# GUARD additional digits.
 | 
						|
#
 | 
						|
# The default 8 is more than enough so no more than 1 correction step
 | 
						|
# was ever needed for all inputs tried through 2.5 billion digits. In
 | 
						|
# fact, I believe 3 guard digits are always enough - but the proof is
 | 
						|
# very involved, so better safe than sorry.
 | 
						|
#
 | 
						|
# Short course:
 | 
						|
#
 | 
						|
# If prec is the decimal precision in effect, and we're rounding down,
 | 
						|
# the result of an operation is exactly equal to the infinitely precise
 | 
						|
# result times 1-e for some real e with 0 <= e < 10**(1-prec). In
 | 
						|
#
 | 
						|
#     ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD
 | 
						|
#     hi = +n * +recip # unary `+` chops to ctx.prec digits
 | 
						|
#
 | 
						|
# we have 3 visible chopped operations, but there's also a 4th:
 | 
						|
# precomputing a truncated `recip` as part of setup.
 | 
						|
#
 | 
						|
# So the computed product is exactly equal to the true product times
 | 
						|
# (1-e1)*(1-e2)*(1-e3)*(1-e4); since the e's are all very small, an
 | 
						|
# excellent approximation to the second factor is 1-(e1+e2+e3+e4) (the
 | 
						|
# 2nd and higher order terms in the expanded product are too tiny to
 | 
						|
# matter). If they're all as large as possible, that's
 | 
						|
#
 | 
						|
# 1 - 4*10**(1-prec). This, BTW, is all bog-standard FP error analysis.
 | 
						|
#
 | 
						|
# That implies the computed product is within 1 of the true product
 | 
						|
# provided prec >= log10(true_product) + 1.602.
 | 
						|
#
 | 
						|
# Here are telegraphic details, rephrasing the initial condition in
 | 
						|
# equivalent ways, step by step:
 | 
						|
#
 | 
						|
# prod - prod * (1 - 4*10**(1-prec)) <= 1
 | 
						|
# prod - prod + prod * 4*10**(1-prec)) <= 1
 | 
						|
# prod * 4*10**(1-prec)) <= 1
 | 
						|
# 10**(log10(prod)) * 4*10**(1-prec)) <= 1
 | 
						|
# 4*10**(1-prec+log10(prod))) <= 1
 | 
						|
# 10**(1-prec+log10(prod))) <= 1/4
 | 
						|
# 1-prec+log10(prod) <= log10(1/4) = -0.602
 | 
						|
# -prec <= -1.602 - log10(prod)
 | 
						|
# prec >= log10(prod) + 1.602
 | 
						|
#
 | 
						|
# The true product is the same as the true ratio n/p256. By Lemma 2
 | 
						|
# above, n.a - p256.a + 1 is an upper bound on the ceiling of
 | 
						|
# log10(prod). Then 2 is the ceiling of 1.602. so n.a - p256.a + 3 is an
 | 
						|
# upper bound on the right hand side of the inequality. Any prec >= that
 | 
						|
# will work.
 | 
						|
#
 | 
						|
# But since this is just a sketch of a proof ;-), the code uses the
 | 
						|
# empirically tested 8 instead of 3. 5 digits more or less makes no
 | 
						|
# practical difference to speed - these ints are huge. And while
 | 
						|
# increasing GUARD above 3 may not be necessary, every increase cuts the
 | 
						|
# percentage of cases that need a correction at all.
 | 
						|
#
 | 
						|
# On Computing Reciprocals
 | 
						|
# ------------------------
 | 
						|
# In general, the exact reciprocals we compute have over twice as many
 | 
						|
# significant digits as needed. 1/256**i has the same number of
 | 
						|
# significant decimal digits as 5**i. It's a significant waste of RAM
 | 
						|
# to store all those unneeded digits.
 | 
						|
#
 | 
						|
# So we cut exact reciprocals back to the least precision that can
 | 
						|
# be needed so that the error analysis above is valid,
 | 
						|
#
 | 
						|
# [Note: turns out it's very significantly faster to do it this way than
 | 
						|
# to compute  1 / 256**i  directly to the desired precision, because the
 | 
						|
# power method doesn't require division. It's also faster than computing
 | 
						|
# (1/256)**i directly to the desired precision - no material division
 | 
						|
# there, but `compute_powers()` is much smarter about _how_ to compute
 | 
						|
# all the powers needed than repeated applications of `**` - that
 | 
						|
# function invokes `**` for at most the few smallest powers needed.]
 | 
						|
#
 | 
						|
# The hard part is that chopping back to a shorter width occurs
 | 
						|
# _outside_ of `inner`. We can't know then what `prec` `inner()` will
 | 
						|
# need. We have to pick, for each value of `w2`, the largest possible
 | 
						|
# value `prec` can become when `inner()` is working on `w2`.
 | 
						|
#
 | 
						|
# This is the `prec` inner() uses:
 | 
						|
#     max(n.a - p256.a, 0) + GUARD
 | 
						|
# and what setup uses (renaming its `v` to `p256` - same thing):
 | 
						|
#     p256.a + GUARD + 1
 | 
						|
#
 | 
						|
# We need that the second is always at least as large as the first,
 | 
						|
# which is the same as requiring
 | 
						|
#
 | 
						|
#     n.a - 2 * p256.a <= 1
 | 
						|
#
 | 
						|
# What's the largest n can be? n < 255**w = 256**(w2 + (w - w2)). The
 | 
						|
# worst case in this context is when w ix even. and then w = 2*w2, so
 | 
						|
# n < 256**(2*w2) = (256**w2)**2 = p256**2. By Lemma 1, then, n.a
 | 
						|
# is at most p256.a + p256.a + 1.
 | 
						|
#
 | 
						|
# So the most n.a - 2 * p256.a can be is
 | 
						|
# p256.a + p256.a + 1 - 2 * p256.a = 1. QED
 | 
						|
#
 | 
						|
# Note: an earlier version of the code split on floor(e/2) instead of on
 | 
						|
# the ceiling. The worst case then is odd `w`, and a more involved proof
 | 
						|
# was needed to show that adding 4 (instead of 1) may be necessary.
 | 
						|
# Basically because, in that case, n may be up to 256 times larger than
 | 
						|
# p256**2. Curiously enough, by splitting on the ceiling instead,
 | 
						|
# nothing in any proof here actually depends on the output base (256).
 | 
						|
 | 
						|
# Enable for brute-force testing of compute_powers(). This takes about a
 | 
						|
# minute, because it tries millions of cases.
 | 
						|
if 0:
 | 
						|
    def consumer(w, limit, need_hi):
 | 
						|
        seen = set()
 | 
						|
        need = set()
 | 
						|
        def inner(w):
 | 
						|
            if w <= limit:
 | 
						|
                return
 | 
						|
            if w in seen:
 | 
						|
                return
 | 
						|
            seen.add(w)
 | 
						|
            lo = w >> 1
 | 
						|
            hi = w - lo
 | 
						|
            need.add(hi if need_hi else lo)
 | 
						|
            inner(lo)
 | 
						|
            inner(hi)
 | 
						|
        inner(w)
 | 
						|
        exp = compute_powers(w, 1, limit, need_hi=need_hi)
 | 
						|
        assert exp.keys() == need
 | 
						|
 | 
						|
    from itertools import chain
 | 
						|
    for need_hi in (False, True):
 | 
						|
        for limit in (0, 1, 10, 100, 1_000, 10_000, 100_000):
 | 
						|
            for w in chain(range(1, 100_000),
 | 
						|
                           (10**i for i in range(5, 30))):
 | 
						|
                consumer(w, limit, need_hi)
 |