Why are they even named like this?
When I read code, I want to be able to read it…
Is this from a time when space was expensive and you wanted to reduce the space of the source files on the devs PC???
For me (with a native language != english), this made it a lot harder to get into programming in the first place.
And with a bit of namespacing and/or object orientation and usage of dots, it becomes perfectly readable.
There are also camel case and underscores in other languages…
BTW: How on earth should a newcomer know that the letter “n” in that word stands for number without having to google it? The newcomer could even assume that it’s a letter of the word string… And even, if you know that it stands for number, it’s still hard for me to understand what it means in this context… I actually had to google it… But that’s probably some C++ convention I don’t know about, because I don’t program in C++…
How on earth should a newcomer know that the letter “n” in that word stands for number without having to google it?
By looking at the difference between strcpy
and strncpy
. Preferably, though, you should simply learn C before writing C.
The gist of is is that strcpy
takes a null-terminated string and copies it somewhere, while strncpy
takes a zero-terminated string and copies it somewhere but will not write more than n
bytes. strncpy
literally has exactly one more parameter than strcpy
, that being n
, hence the name. If n
is smaller than the string length (as in: distance to first null byte) then you’re bound to have garbage in your destination, and to check for that you have to dereference the pointer strncpy
returns and check if it’s actually null. Yay C error handling.
In retrospect null-terminated strings were a mistake, but so were many other things, at some point you just have to accept that there’s hysterical raisins everywhere.
C is a little older than namespacing and object orientation. C++ wasn’t even a glimmer in Bjarne’s eye when these conventions were laid down.
And yes, having to google it is part of the design. Originally C programmers would have had to read actual manuals about this stuff. Once you learn the names you don’t really forget so it works well enough even now for ubiquitous standard library functions.
And yet, C was an ergonomic revelation to programmers of the time. Now it’s the arcane grandpa that most youngsters don’t put up with.
Did you know that in the first version of php, each function name would be hashed to lookup the code to run it? And the hashing algorithm was: the first letter. So all the functions started with a different letter.
It’s not. PHP used to use the function length as hash buckets, so by having evenly distributed lengths the execution time was faster. No idea where GP came up with that.
I recently held a science slam about this topic! It’s a mix of the first computer scientists being mathematicians, who love their abbreviations, and limited screen size, memory and file size. It’s a trend in computing that has been well justified in the past, but has been making it harder for people to work together. And the need to use abbreviations has completely gone with the age of auto completion and language servers.
mathematicians, who love their abbreviations
Man, I hate that so much. I swear this was half the reason I struggled with maths and physics, that these guys need to write this:
Rather than this:
At some point, they even collectively decided that not having to write a multiplication dot is more important than being able to use more than a single letter for your variables. Just what the fuck?
Using full names like that might be fine for explaining a physical rule, or stating the final result of some calculation - but it certainly would be cumbersome and difficult for actually carrying out the calculations. In many cases we already fill pages with algebra showing how things can be related and rearranged to arrive at new results. That kind of work would be intractable with full word names for the variables, partially because you’d be constantly spilling off the end of the page trying to write the steps; but also because having all that stuff would actually obfuscate what you are trying to do - which is algebra. And during that process, the meanings and values of the pronumerals is not as important has how they interact with each other. So the names are just a distraction.
For setting up an equation, and for stating the final result, the meanings of the variables are very important; but during the process of manipulating the equations to get the result you want the meanings of the letters are often ignored. You only need to know that it is something that can be multiplied, or inverted, or subtracted, or whatever. Eg. suppose I want to rearrange to get the velocity. I don’t care that I’m dividing both sides by the air density times the drag coefficient and the area… I’m just dividing ρCA, which is an algebraic blob whose interpretation can be saved for some other time.
I would have quit math if I had to do algebra with names instead of letters.
Thing is, you usually define all your variables. At least we do in engineering (of physical variety, rather than software).
Mostly because we can’t expect everyone reading the calculation to know, and that not everyone uses the same symbols.
Not explaining each variable is bad practice, other than for very simple things. (I do expect everyone and their dog reading a process eng calc to know PV=nRT, at a minimum).
Just like (in my opinion) not defining industry specific abbreviations is also bad practice.
Mathematicians don’t do this? Shame on them.
Try to write the above with pen and ink and then tell me if you can read it back yourself.
Single letters is not a good system but it was the less bad one.
It’s been really holding me back in learning coding. I felt pretty comfortable at first learning javascript, but as I got further the code was increasingly hard to look back to and understand, to the point I had to spend a lot of time understanding my own code.
Does it truely matter after the code has been compiled if it has more full words or not?
It matters as soon as a requirement change comes in and you have to change something. Writing a dirty ass incomprehensible, but working piece of code is ok, as long as no one touches it again.
But as soon as code has to be reworked, worked on together by multiple people, or you just want to understand what you did 2 weeks earlier, code readability becomes important.
I like Uncle Bobs Clean Code (with a grain of salt) for a general idea of what such an approach to make code readable could look like. However, it is controversial and if overdone, can achieve the opposite. I like it as a starting point though.
It’s from a time keyboards were so hard that you needed to do push ups on your finger tips if you wanted to endure a 9 to 5 programming job.
The reason people love IBM Model Ms nowadays is because the springs have been worn in now and can easily be pressed without additional training.