High Level
I finally found a well-written no-bullshit book about CS. It says, among other things: There is no need to define a representation of the values False and True in terms of values of some other type. Conceptually, values of type Bool are simply (denoted by) the expressions False and True. Of course, the computer’s internal representation of Bool and all other data is in terms of bits, but we don’t need to know any of the details of how that works to write programs....