The title would probably be confusing, but I could not make it better than this. I noticed that most programming languages are limited to the alphanumerical set along with the special characters present in a general keyboard. I wondered if this posed a barrier for developers on what characters they were limited to program in, or if it was intentional from the start that these keys would be the most optimal characters for a program to be coded in by a human and was later adopted as a standard for every user. Basically, are the modern keyboards built around programming languages or are programming languages built around these keyboards?

  • themoonisacheese@sh.itjust.works
    link
    fedilink
    arrow-up
    50
    ·
    1 year ago

    Pretty much all programming languages in use today have been invented decades after the keyboard.

    The keyboard was originally invented for typewriters, but since they don’t interact with anything and are self-contained systems, typewriters don’t need their keys standardized. However, the keyboard converged as HP, Brother et al. brought the typewriter forward, especially near the end when typewriters were keyboards connected to printers.

    Early computers weren’t programmed with keyboards at all and used punch cards to be programmed. As you move forwards with the invention of the computer as we know it today, you find out that the modern keyboard was invented to be able to type most characters in use at the time (a classic example is that the @ sign was used by tellers on tills, to mark stuff like 2 units @ $5 per), and actual standards body such as iso and ANSI simply implemented that.

    Early personal computers were not powerful enough to compile programs quickly(and compilers were very expensive), so for a while all programming was done either in assembly, which requires very few special characters, or in higher lever languages by people who could pay to get their programs compiled. These higher lever languages, such as B (precursor to C), were often written right on paper by hand, or by typewriter. After this point, home compilers became more and more accessible (the internet basically guaranteed you could get a compiler if needed), and this was the last point in history you could design a language with characters not present on a standard keyboard (you could still have a nonstandard typewriter, or write them by hand for example). After this and the explosion of home computing, people designed their own languages that they were able to use themselves, using their own keyboards.

    Worth noting that there exist languages today that you cannot type, pretty much at all. They are however largely considered esolangs (esoteric languages) and most of their use is relegated to the practice of code-golf, in which a programmer tries to accomplish a task using the fewest bytes. Since the characters on a keyboard make up a fraction of all characters available in Unicode or even just ASCII, these languages try to increase the number of things you can do with a single characters by disregarding the fact you can’t type most of them, lowering the amount of bytes one needs to do a task.

    An example of this is O5AB1E (pronounced “osable”), which has over 250 single-character commands, such as Δ which “repeats code until ‘a’ doesn’t change”.

    An example (that you can’t type) is: yā<ã.Δ¹sŸèOQ

    This program gets an input n and a string of numbers, then outputs the first run in the string that sums to n. The program is explained here

    • jadero
      link
      fedilink
      arrow-up
      16
      ·
      edit-2
      1 year ago

      I’d just like to take a moment to commend you on an absolutely exquisite answer.

      Given time, there is a slim chance I could have covered the same ground and no chance that it would have been as clear and concise.

    • snoweMA
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      An example (that you can’t type) is: yā<ã.Δ¹sŸèOQ

      I think the only thing here you can’t type is the superscript. The rest is easily typeable, at least on a Mac or a smartphone.

  • fubo@lemmy.world
    link
    fedilink
    arrow-up
    21
    ·
    edit-2
    1 year ago

    C supports alternate ways of typing some of its punctuation, for programmers whose keyboards didn’t support them all. For example, if you can’t type [ ] you can use ??( ??) instead. (There are other ones that use angle brackets, but I can’t type them here because Lemmy escapes them incorrectly. Irony.)

    https://en.wikipedia.org/wiki/Digraphs_and_trigraphs#C


    The usual example of a programming language using especially unusual characters is APL, where all built-in functions are all represented by single characters, mostly drawn from mathematical notations, Greek alphabet, and so on. For example, is the “sort” function.

      • fubo@lemmy.world
        link
        fedilink
        arrow-up
        5
        ·
        1 year ago

        The core of that Life expression works by taking the array of cells and shifting it in the eight different directions, then summing those arrays to get the population counts.

        I tried translating this into Python — but I’ve never written numpy code before, so this is probably less efficient than it could be. But it does work and you can see a glider move through a few generations.

        The array-shifting logic is in the populations function, with np.roll being the equivalent of the APL rotate operation (written as and in the original).

        import numpy as np
        
        def alive(popu, cell):
            return int(popu == 3 or (popu == 4 and cell))
        alive = np.frompyfunc(alive, 2, 1)
        
        def populations(grid):
            return np.array(
                    [ np.roll(r1, shift, 1)
                        for shift in [-1, 0, 1]
                        for r1 in (np.roll(grid, shift, 0) for shift in [-1, 0, 1]) ]
                    ).sum(axis=0)
        
        def nextgen(grid):
            return alive(populations(grid), grid)
        
        grid = np.array([[0, 0, 0, 0, 0, 0],
                         [0, 0, 1, 0, 0, 0],
                         [0, 0, 0, 1, 0, 0],
                         [0, 1, 1, 1, 0, 0],
                         [0, 0, 0, 0, 0, 0],
                         [0, 0, 0, 0, 0, 0]])
        
        for gens in range(5):
            print(grid)
            grid = nextgen(grid)
        print(grid)
        
  • marcos@lemmy.world
    link
    fedilink
    arrow-up
    22
    arrow-down
    1
    ·
    1 year ago

    It was a noisy co-evolution where both languages and keyboards keep changing to best fit each other.

    While the letter layout of our keyboards is heavily influenced by typewriters, the set of symbols on them changed a lot with times, and for some times there was a lot of diversity on them. Our current design is as much a result of languages like MPL never really popularizing as much as the lack of adoption of those languages is a result of the popular keyboard designs.

    • Vashti@feddit.uk
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      Yeah. 7-bit ASCII goes back to what, the 60s? But computers still used different encodings and so different keyboards for A While.

  • Turun@feddit.de
    link
    fedilink
    arrow-up
    17
    ·
    1 year ago

    Not only do the Languages orient themselves to fit to the keyboard, they pretty much only consider the English keyboard layout.

  • jflorez@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    16
    ·
    1 year ago

    Modern keyboards are an evolution of the old typewriter that existed before computers as we know them and programming as we know it existed, so I would say keyboard came first

      • mo_ztt ✅@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        1 year ago

        He’s not completely, though. @marcos had it right about co-evolution – leaving aside any issues of internationalization, the layout of letters came from typewriters, but the layout of punctuation available was different on different computers for a lot of the early history of programming. Some of the more extreme examples were the Space Cadet Keyboard used at MIT, and APL which more or less required you to use an APL-specific keyboard in order to be able to access all the special symbols that are part of APL. Here’s an APL program:

        ⎕CR 'ProveNonPrime'
        Z←ProveNonPrime R
        ⍝Show all factors of an integer R - except 1 and the number itself,
        ⍝ i.e., prove Non-Prime. String 'prime' is returned for a Prime integer.
        Z←(0=(⍳R)|R)/⍳R  ⍝ Determine all factors for integer R, store into Z
        Z←(~(Z∊1,R))/Z   ⍝ Delete 1 and the number as factors for the number from Z.
        →(0=⍴Z)/ProveNonPrimeIsPrime               ⍝ If result has zero shape, it has no other factors and is therefore prime
        Z←R,(⊂" factors(except 1) "),(⊂Z),⎕TCNL  ⍝ Show the number R, its factors(except 1,itself), and a new line char0  ⍝ Done with function if non-prime
        ProveNonPrimeIsPrime: Z←R,(⊂" prime"),⎕TCNL  ⍝ function branches here if number was prime
        

        Things became much more standardized when the IBM PC’s keyboard became the norm, and were formalized in 1995 with ISO 9995. Then once it stabilized there was a strong incentive for both language designers and keyboard makers to stick with what everyone was used to so they could keep working with the other. But it wasn’t always that way.

        Edit: Here’s what things looked like on an IBM 3276:

        (Full size image)

      • CrunchyBoy@lemmy.world
        link
        fedilink
        arrow-up
        3
        arrow-down
        3
        ·
        edit-2
        1 year ago

        Fun fact: the standard qwerty layout was made to slow typewriter typing down by putting common keys off the home row and apart from each other. This was done to prevent the little key arm thingies from colliding and jamming when typing quickly.

        EDIT: Apparently this is not a fact

  • OOFshoot@beehaw.org
    link
    fedilink
    arrow-up
    15
    ·
    1 year ago

    Programming languages are build around the standard keyboard. Keyboards had most of the symbols you’re thinking of from their typewriter days. You can see most of the special characters in these small typewriters from the mid 1900s.

    https://dealdashreviewed.com/wp-content/uploads/2015/02/typerwriter.jpg

    https://3.bp.blogspot.com/-gdj1kOXgkTY/UqYxNGAxBBI/AAAAAAAA8T0/DTeg3C_ydXM/s1600/71-6zsvcleL._SL1500_.jpg

    With things like electric Wheel Writer typewriters, adding extra keys and symbols were less of a complexity issue and you started to see a few more extra symbols.

    https://www.imagine41.com/wp-content/uploads/2016/07/ibm_wheelwriter_2500_002_1.jpg

    Recognize that there never has been a hard standard for layouts and symbols, just the industry copying and converging on systems that became popular.

  • Die4Ever
    link
    fedilink
    arrow-up
    12
    ·
    edit-2
    1 year ago

    I won’t consider any keyboard to be designed for programming unless it has dedicated keys for characters like {}() &lt; >_+| &amp; !*:" without needing to hold shift for them (Lemmy seems to be improperly escaping my less-than sign and ampersand)

  • BradleyUffner@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    2
    ·
    1 year ago

    Some of the “toy” languages used for code-golf can use some very non standard characters. Saying “all programming languages” is going to be very restrictive.

    • jvisick
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      That’s a little pedantic, don’t you think?

      • BradleyUffner@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Ohh very much so. This wasn’t meant as a criticism of OP’s question. I was just trying to make sure they were aware of all the niche programming languages out there, especially the ones that use some rather exotic character sets.

    • sylowosa@lemm.eeOP
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Fixed ‘all’. I tried to acknowledge that there may be other languages that may not follow this rule, but I seem to have not noticed this mistake.