I’ve been using neovim for years (and the vim family for decades), and I guess with LSP it’s pretty much an IDE these days.
I’ve been using neovim for years (and the vim family for decades), and I guess with LSP it’s pretty much an IDE these days.
It’s also likely a bit of cost-benefit analysis for self-hosting vs using a managed service.
Codeberg would be more in line with Mozilla’s ideals IMO, but GitHub is a pragmatic choice anyway.
Phabricator was an alternative for a development platform of sorts; development ceased in 2021. They’re still running here and there, but I expect them to be in the process of being deprecated.
Yes I’m being sarcastic, but I also think utf-8 is plaintext these days. I really can’t spell my name in US ASCII. Like the other commenter here went into more detail on, it has its history, but isn’t suited for today’s international computer users.
It’s also some surprise internal representation as utf-16; that’s at least still in the realm of Unicode. Would also expect there’s utf-32 still floating around somewhere, but I couldn’t tell you where.
And is mysql still doing that thing with utf8
as a noob trap and utf8_for_real_we_mean_it_this_time_honest
or whatever they called it as normal utf8?
Yes, I am joking. We probably could do something like the old iso-646 or whatever it was that swapped letters depending on locale (or equivalent), but it’s not something we want to return to.
It’s also not something we’re entirely free of: Even though it’s mostly gone, apparently Bulgarian locales do something interesting with Cyrillic characters. cf https://tonsky.me/blog/unicode/
To unjerk, as it were, it was a thing. So on old systems they’d do stuff like represent æøå with the same code points as {|}
. Curly brace languages must have looked pretty weird back then:)
Jess. Ai’m still lukking får the ekvivalent åv /r/JuropijenSpelling her ån lemmi. Fæntæstikk søbreddit vitsj æbsolutli nids lemmi representeysjen.
No, I’m pretty sure the weird o with the leg is in basic ASCII. It’s also missing Latin characters like Æ. It’s a very weird standard.
Q. P is a common character across languages. But Q is mostly unused, at least outside the romance languages who appear to spell K that way. But that can be solved by letting the characters have the same code point, and rendering it as K in most regions, and Q in France. I can’t imagine any problems arising from that. :)
I’m not entirely sure here, but you are aware you’re in a humour community, yeah?
Neovim developer got sidetracked configuring their reply plugin
It’s a joke because it includes useless letters nobody needs, like that weird o with the leg, and a rich set of field and record separating characters that are almost completely forgotten, etc, but not normal letters used in everyday language >:(
With ASCII æs the åriginal sin. Can’t even spell my name with that joke of an encoding >:(
Isn’t that sort of just the cost of doing business in C? It’s a sparse language, so it falls to the programmer to cobble together more.
I do also think the concrete example of emails should be taken as a stand-in. Errors like swapping a parameter for an email application is likely not very harmful and detected early given the volume of email that exists. But in other, less fault-tolerant applications it becomes a lot more valuable.
It is pretty funny that C’s type system can be described pretty differently based on the speaker’s experience. The parable of the Blub language comes to mind.
Parsing is a way of “validating early”. You either get a successful parse and the program continues working on known-good data with that knowledge encoded in the type system, or you handle incorrect data as soon as it’s encountered.
I used Ratpoison for well over a decade, and only replaced it with sway once I had a new machine and figured it was time to try Wayland. Apparently that’s some 4-5 years ago already.
I feel I gotta point out it’s a pretty funny example—email comes up so frequently as a thing that you’re recommended to neither parse nor validate, just try to send an email to the address and see if it works. If you need to know that it was received successfully, a link to click is the general method.
But “parse, don’t validate” is still a generally good idea, no matter the example used. :)
That depends on when it appears. Some tasks kind of have to feel instantaneous, and there might be a pretty slim margin between okay and frustrating.
But yeah, that’s the kind of savings that mostly matter on the scale of regional or national grid planning.
Yeah, the author seems to lean too hard into the “programming is electronics” model, where the opposing end is “programming is math and formal logic”; most of us take some mixed view. And most of us have higher correctness requirements than what a reasonable effort in memory unsafe languages like C and C++ gives us, so we trade away some machine efficiency. In the authors parlance, most of us aren’t interested in the demoscene circlejerk; we need to make tradeoffs between maintainability and everything else. Write-once code isn’t good enough.
There have been attempts at establishing a third pole of “promptgramming is natural language” or whatever ever since COBOL promised programming in plain English, but the ambiguity of natural language when used to encode a business logic machine means that a “sufficiently advanced compiler” will have to be extremely advanced, on the order of including the manager and the entire engineering methodology.