When converting an integer to a bit array, the length of the bit array is the number of bits in the integer type, and the
bit array's value is the binary representation. The most significant bit of the integer becomes the first bit of the array.
When converting a binary type of length n to a bit array, the length of the array is n * 8 bits. The first 8 bits of the bit array become the first byte of the binary value. The most significant bit of the binary
value becomes the first bit in the array. The next 8 bits of the bit array become the second byte of the binary value, and
so on.
When converting a character data type of length n to a bit array, the length of the array is n bits. Each character must be either '0' or '1' and the corresponding bit of the array is assigned the value 0 or 1.
When converting a bit array to an integer data type, the bit array's binary value is interpreted according to the storage
format of the integer type, using the most significant bit first.
When converting a bit array to a binary, the first 8 bits of the array become the first byte of the binary value. The first
bit of the array becomes the most significant bit of the binary value. The next 8 bits are used as the second byte, and so
on. If the length of the bit array is not a multiple of 8, then extra zeroes are used to fill the least significant bits of
the last byte of the binary value.
When converting a bit array of length n bits to a character data type, the length of the result is n characters. Each character in the result is either '0' or '1', corresponding to the bit in the array.