Alt text:

Transcendence meme template

function main() {…}

int main() {…}

void main() {…}

U0 main() {…}

/* HolyC example */
U0 Main()
{
  U8 *message = "hello world";
  "%s\n",message;
}
Main;
    • Ephera@lemmy.ml
      link
      fedilink
      English
      arrow-up
      6
      ·
      3 days ago

      What really frustrates me about that, is that someone put in a lot of effort to be able to write these things out using proper words, but it still isn’t really more readable.

      Like, sure, unsigned is very obvious. But short, int, long and long long don’t really tell you anything except “this can fit more or less data”. That same concept can be expressed with a growing number, i.e. i16, i32 and i64.

      And when someone actually needs to know how much data fits into each type, well, then the latter approach is just better, because it tells you right on the tin.

      • labsin@sh.itjust.works
        link
        fedilink
        arrow-up
        9
        ·
        3 days ago

        In c they do indeed just mean shorter and longer int as the size of the int is defined by the compiler and target and originally represented the hardware.

        There are types like int32_t or int_least16_t.

        • Ephera@lemmy.ml
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 days ago

          Huh, so if you don’t opt for these more specific number types, then your program will explode sooner or later, depending on the architecture it’s being run on…?

          I guess, times were different back when C got created, with register size still much more in flux. But yeah, from today’s perspective, that seems terrifying. 😅

          • fiqusonnick@lemmy.blahaj.zone
            link
            fedilink
            arrow-up
            2
            ·
            21 hours ago

            The C standard for different ints is absolutely cursed, even after C99 tried to normalize it. The only requirement is that sizeof(char) <= sizeof(short) <= sizeof(int) <= sizeof(long) <= sizeof(long long) and sizeof(char) == 1. Mind you they don’t define what size a byte is so you technically can have an architecture where all of those are 64 bits. Oh and for that same reason exact-size types (int32_t, uint16_t etc) are not guaranteed to be defined

            Fuck