

That’s quite a strong table, holding 11 people
That’s quite a strong table, holding 11 people
A “server” is just a remote computer “serving” you stuff, after all. Although, if you have stuff you would have trouble setting up again from scratch, I’d recommend you look into making at least these parts of your setup repeatable, be it something fancy ala Ansible, or even just a couple of bash scripts to install the correct packages and backing up your configs.
Once you’re in this mindset and take this approach by default, changing machines becomes a lot less daunting in general. A new personal machine takes me about an hour to setup, preparing the USB included.
If it’s stuff you don’t care about losing, ignore everything I just said. But if you do care about it, I’d slowly start by giving from the most to least critical parts. There’s no better time to do it than when things are working well haha!
58% goes to fundraising, administrative and technological costs. The rest has some money going towards, but no limited to, other programs.
Only thing I can find in their financials that would maybe qualify as “random outreach” would be “awards and grants”, at 26mil last year out of 185mil revenue, or 14%.
https://meta.m.wikimedia.org/wiki/Grants:Programs/Wikimedia_Community_Fund
As far as I can tell, it’s not particularly random.
Maybe I’m missing something?
Tramp is more featured, but if all one cares about is being able to edit remote files using a local editor, vim can edit remote files with scp too: scp://user@server[:port]//remote/file.txt
I tried tramp-mode at some point, but I seem to remember some gotchas with LSP and pretty bleh latency, which didn’t make it all that useful to me… But I admittedly didn’t spend much time in emacs land.
Eh, I’m about the same age as OP, I don’t have to get to 50 to know that I’d take my parents’ economic context over the two crashes. The rest… For many reasons, if medicine does some miraculous leap forward by then, maybe I’ll still wish I got a lot more left to go by then.
Really bigger updates obviously require a major version bump to signify to users that there is potential stability or breakage issues expected.
If your software is following semver, not necessarily. It only requires a major version bump if a change is breaking backwards compatibility. You can have very big minor releases and tiny major releases.
there was more time for people to run pre-release versions if they are adventurous and thus there is better testing
Again, by experience, this is assuming a lot.
From experience shipping releases, “bigger updates” and “more tested” are more or less antithetical. The testing surface area tends to grow exponentially with the amount of features you ship with a given release, to the point I tend to see small, regular releases, as a better sign of stability.
I’d love to share your optimism, especially regarding that last sentence. As long as Google controls the most popular web browser out there, I don’t see the arms race ever stopping, they’ll just come up with something else. It wouldn’t be the first time they push towards something nobody asked for that only benefits themselves.
I do connect to VMs and containers all the time, I just don’t see a reason not to speed myself up on my own machines because of it. To me, the downside of typing an alias on a machine that doesn’t have it once in a while, is much less than having to type everything out or searching my shell history for longer commands every single time. My shell configs are in a dotfiles repo I can clone to new personal/work machines easily, and I have an alias to rsync some key parts to VMs if needed. Containers, I just always assume I don’t have access to anything but builtins. I guess if you don’t do the majority of your work on a local shell, it may indeed not be worth it.
I’d rather optimize for the 99% case, which is me getting shit done on my machine, than refuse to use convenient stuff for the sake of maybe not forgetting a command I can perfectly just look up if I do legitimately happen to forget about it. If I’m on a remote, I already don’t have access to all my usual software anyway, what’s a couple more aliases? To me this sounds like purposefully deciding to slow yourself down cutting paper with a knife all the time cause you may not have access to scissors when you happen to sit at someone else’s desk.
That’s not “self hosting” related tho lol
It desperately needs interface types if we ever hope to make it a serious contender for general purpose web development. The IO overhead of having to interface with JS to use any web API is itself pretty slow, and is limiting a lot of usecases.
Considering the community we are on, I assumed the criticism was more about the privacy problems surrounding the engine and browser security model than the quality of the language itself. If that was the intent, I mean… Yeah, its weak typing is a fucking mess.
The stuff like Flash, Java applets and Silverlight it eventually replaced were arguably even worse. There’s a legitimate need to run client-side code at times, IMHO the mistake was making it so permissive by default. Blaming the language for the bad browser security model is kind of throwing away the baby with the bathwater.
Is it solid wood or engineered? Some very soft variety of wood? 17 years is extremely short…
As of 2021, the US spent 16.6% of its gross GDP ($23.59 billions) on healthcare expenditures. The very next was Germany, at 12.7% of its $4.28 billion GDP. The US is spending more per-capita than any other OECD country on healthcare, it’s just not made visible by looking at the number on your tax report. You’re still collectively paying for it one way or another.
But hey, yay, low taxes. Good for you, I guess?
Considering how little we actually know, how much we are still figuring out today, how wrong we once were, and most definitely still are on many things, about said nature, the naturalistic argument is IMHO rather weak. The argument silently assumes too many things, at least with our current knowledge - that human beings do actually have an inherent nature, that said nature is uniform enough across the whole species to make that generalization, that said nature is inevitable and can’t be evolved past or rationalized against, that it always was the case and will always be, etc.
Ah, well that’s what almost always ends up happening, doesn’t it… The only thing that legitimately trickles down in this fucking system is costs to consumers lol
I’m not saying the middle ground doesn’t exist, but that said middle ground visibly doesn’t cause enough damage to businesses’ bottom line, leading to companies having zero incentive to “fix” it. It just becomes part of the cost of doing business. I sure as hell won’t blame programmers for business decisions.
Interesting. I interpreted this definition more like an oval vs. circle distinction. The vast majority of ovals aren’t circles, but circles are a subset of ovals.