Why did K&R not define data type lengths better?

257 Views Asked by At

Every modern programming language I've encountered has well-defined data types. Java, for example, has an int that is exactly 32 bits and a long that is exactly 64 bits. This is not implementation specific but built into the spec itself. C, on the other hand, has a short that is at least 16 bits, an int that is at least 16 bits, and a long that is at least 32 bits. Because it is defined as at least, though, they can be much larger, which results in serious portability issues.

I have never understood this. I have always chalked it up to being the result of differing hardware standards in the 1970's, but this doesn't actually make sense to me. Even if there were a lot of computers that couldn't handle 64 bit data types, that still doesn't excuse not defining a short as necessarily 16 bits and an int as necessarily 32 bits. From the standards we already see 32 bits as a minimum requirement for any hardware.

This proliferation of standards has resulted in advice like the following, taken from the article How to C (as of 2016):

If you find yourself typing char or int or short or long or unsigned into new code, you’re doing it wrong.

For modern programs, you should #include <stdint.h>then use standard types.

For more details, see the stdint.h specification.

The common standard types are:

  • int8_t, int16_t, int32_t, int64_t — signed integers
  • uint8_t, uint16_t, uint32_t, uint64_t — unsigned integers
  • float — standard 32-bit floating point
  • double - standard 64-bit floating point

Notice we don’t have char anymore. char is actually misnamed and misused in C.

Developers routinely abuse char to mean “byte” even when they are doing unsigned byte manipulations. It’s much cleaner to use uint8_t to mean single a unsigned-byte/octet-value and uint8_t * to mean sequence-of-unsigned-byte/octet-values.

I've also noticed quite a bit of advice recommending the use of size_t as the standard go to int replacement, as in Modern C by Jens Gustedt which uses a size_t as the control variable in a for loop in the very first example of the book.

My question is two fold:

1. Why did K&R 2 not define data types more definitely in 1978?

2. What, in 2018, are we supposed to do with our data type choices, or is it all just style and convention and highly opinionated?

0

There are 0 best solutions below