Fix warnings about using char as an array subscript. This is not portable

since char is signed on some platforms and unsigned on others.
This commit is contained in:
Neal Norwitz 2008-03-27 04:40:50 +00:00
parent 4ebd46a02d
commit 231346e23f
4 changed files with 22 additions and 22 deletions

View file

@ -1397,7 +1397,7 @@ long_from_binary_base(char **str, int base)
n >>= 1;
/* n <- total # of bits needed, while setting p to end-of-string */
n = 0;
while (_PyLong_DigitValue[Py_CHARMASK(*p)] < base)
while (_PyLong_DigitValue[(unsigned)Py_CHARMASK(*p)] < base)
++p;
*str = p;
/* n <- # of Python digits needed, = ceiling(n/PyLong_SHIFT). */
@ -1418,7 +1418,7 @@ long_from_binary_base(char **str, int base)
bits_in_accum = 0;
pdigit = z->ob_digit;
while (--p >= start) {
int k = _PyLong_DigitValue[Py_CHARMASK(*p)];
int k = _PyLong_DigitValue[(unsigned)Py_CHARMASK(*p)];
assert(k >= 0 && k < base);
accum |= (twodigits)(k << bits_in_accum);
bits_in_accum += bits_per_char;
@ -1609,7 +1609,7 @@ digit beyond the first.
/* Find length of the string of numeric characters. */
scan = str;
while (_PyLong_DigitValue[Py_CHARMASK(*scan)] < base)
while (_PyLong_DigitValue[(unsigned)Py_CHARMASK(*scan)] < base)
++scan;
/* Create a long object that can contain the largest possible
@ -1635,10 +1635,10 @@ digit beyond the first.
/* Work ;-) */
while (str < scan) {
/* grab up to convwidth digits from the input string */
c = (digit)_PyLong_DigitValue[Py_CHARMASK(*str++)];
c = (digit)_PyLong_DigitValue[(unsigned)Py_CHARMASK(*str++)];
for (i = 1; i < convwidth && str != scan; ++i, ++str) {
c = (twodigits)(c * base +
_PyLong_DigitValue[Py_CHARMASK(*str)]);
_PyLong_DigitValue[(unsigned)Py_CHARMASK(*str)]);
assert(c < PyLong_BASE);
}

View file

@ -480,13 +480,13 @@ PyObject *PyUnicode_FromStringAndSize(const char *u, Py_ssize_t size)
/* Single characters are shared when using this constructor.
Restrict to ASCII, since the input must be UTF-8. */
if (size == 1 && Py_CHARMASK(*u) < 128) {
unicode = unicode_latin1[Py_CHARMASK(*u)];
unicode = unicode_latin1[(unsigned)Py_CHARMASK(*u)];
if (!unicode) {
unicode = _PyUnicode_New(1);
if (!unicode)
return NULL;
unicode->str[0] = Py_CHARMASK(*u);
unicode_latin1[Py_CHARMASK(*u)] = unicode;
unicode_latin1[(unsigned)Py_CHARMASK(*u)] = unicode;
}
Py_INCREF(unicode);
return (PyObject *)unicode;