I got the hardware survey on my Windows PC, but not on my Steamdeck. So I wonder if there is only 1 survey per user, and most people don’t use a steamdeck exclusively?
I made LASIM! https://github.com/CMahaff/lasim
I currently have 3 accounts (big shock):
I got the hardware survey on my Windows PC, but not on my Steamdeck. So I wonder if there is only 1 survey per user, and most people don’t use a steamdeck exclusively?
Out of curiosity, what content are you looking for? Discovery on Lemmy can be a problem, but sometimes the communities are there and even active, just buried.
But may I also suggest searching by Top Day/12-hour/6-hour to see the most active posts. Lemmy’s scaled algorithm still doesn’t get it quite right IMO.
The CEO said they were going to add pay-walled subreddits at an earnings call.
So… Yep.
I know for me, at least with gnome, toggling between performance, balanced, and battery saver modes dramatically changes my battery life on Ubuntu, so I have to toggle it manually to not drain my battery life if it’s mostly sitting there. I don’t know if Mint is the same, but just throwing out the “obvious” for anyone else running Linux on a laptop.
Found a blog post that gives a quick overview of how to do git via email in general: https://peter.eisentraut.org/blog/2023/05/09/how-to-submit-a-patch-by-email-2023-edition
So at least from my understanding you’d make your changes, email the contents of the patch to the maintainer, and then they’d apply it on their side, do code review, email you comments, etc. until it was in an acceptable state.
There’s also the full kernel development wiki that goes into all the specifics: https://www.kernel.org/doc/html/v4.16/process/howto.html
(I never got through the whole thing)
I’ll also throw out: aging infrastructure, build systems, coding practices, etc.
I looked into contributing to the kernel - it’s already an uphill battle to understand such a large, complex piece of software written almost entirely in C - but then you also need to subscribe to busy mailing lists and contribute code via email, something I’ve never done at 30 and I’m betting most of the younger generation doesn’t even know is possible. I know it “works” but I’m really doubting it’s the most efficient way to be doing things in 2024 - there’s a reason so many infrastructure tools have been developed over the years.
The barriers to entry for a lot of projects is way too high, and IMO a lot of existing “grey” maintainers, somewhat understandably, have no interest in changing their processes after so much time. But if you make it too hard to contribute, no one will bother.
Out of curiosity, what switch are you using for your setup?
Last time I looked, I struggled to find any brand of “home tier” router / switch that supported things like configuring vlans, etc.
Maybe I am not thinking of the access control capability of VLANs correctly (I am thinking in terms of port based iptables: port X has only incoming+established and no outgoing for example).
I think of it like this: grouping several physical switch ports together into a private network, effectively like each group of ports is it’s own isolated switch. I assume there are routers which allows you to assign vlans to different Wi-Fi access points as well, so it doesn’t need to be literally physical.
Obviously the benefits of vlans over something actually physical is that you can have as many as you like, and there are ways to trunk the data if one client needs access to multiple vlans at once.
In your setup, you may or may not benefit, organizationally. Obviously other commenters have pointed out some of the security benefits. If you were using vlans I think you’d have at a minimum a private and public vlan, separating out the items that don’t need Internet access from the Internet at all. Your server would probably need access to both vlans in that scenario. But certainly as you say, you can probably accomplish a lot of this without vlans, if you can aggressively setup your firewall rules. The benefit of vlans is you would only really need to setup firewall rules on whatever vlan(s) have Internet access.
I loved the original Hades, but I played it after it left Early Access.
It’s going to be really hard to resist jumping in early with Hades II.
So maybe you’ve heard something I haven’t, according to this timeline, there was only 2-3 minutes between when the ship issued a mayday and the bridge collapsed: https://www.cbsnews.com/news/francis-scott-key-bridge-collapse-timeline-911-call-dali-cargo-ship-mayday-maps-construction-worker-recovery/
It sounded like there was 1 police officer already stationed on either end of the bridge, so thats the only reason they were even able to close the bridge before the collapse.
In the time it took them to do that, I can’t see how there would have been time to warn them physically (it’s like a 2 mile bridge). From the article, it sounded like there was confusion about if a crew was even on the bridge. I also don’t know how often / what mechanism police can use to directly contact crews, if there even is one.
I ran into the same thing. I’ve always just worked around it, but I believe I did find the solution at one point (can’t find the link now).
But if I am remembering right, I believe you need to manually create a bridge between the two networks - by default it isolates the VMs from TrueNAS itself for security reasons.
Sorry I can’t link the exact fix right now, but hopefully this will help you Google the post I found on the subject.
I dunno man, I think that the fact she teaches high school kids specifically, who by now all know about it, means that she has no hope of being an effective teacher at this point. It’s a massive distraction, as unfair as that is.
She had to have known this was a possibility when she decided to start an onlyfans - there’s almost nowhere in the country where you won’t get fired as a teacher for that, progressive or conservative states alike. Society just isn’t there yet.
I ran into this exact situation at work - though for me it was more the case that getting approvals for new software / installing new dependencies in our system is a massive pain.
So I went with Python since it’s already installed on basically any Linux system. It was fine - I mean Python is a good language and can certainly handle string processing and data manipulation with relative ease.
I still think the Python docs are pretty bad, and I wasn’t thrilled with the options for calling a subprocess in Python - they all felt kinda clunky, though I was barred from using the newest versions since I had to run an older version of Python.
But I ultimately got something that worked and it was certainly better executed / shorter than the bash equivalent it was replacing.
To expand on this a bit, git pull
under the hood is basically a shortcut for git fetch
(get the remote repository’s state) and git merge origin/main main
(merge to remote changes to your local branch, which for you is always main).
When you have no local changes, this process just “makes a line” in your commit history (see git log --graph --decorate
), but when you have local changes and the remote has changed too, it has to put those together into a merge commit - think a diamond shape with the common ancestor at the bottom, the remote changes on one side, your changes on the other side, and the merge of the two at the top.
Like the above comment says, normally this process is clarified at the command line - VSCode must be handling it automatically if there are no code conflicts.
You offered a lot of suggestions, and I’m sure people will disagree over the specifics, but I think your overall point is excellent and not talked about enough. I wonder if anyone has ever even attempted a survey on the ages of maintainers/contributors? I bet it’s skewing older fast.
Nothing wrong with that of course, especially given the project’s age, complexity, and being written in C - but you’re right, at some point you have to attract new talent - people can’t maintain forever.
I’m a 29 year old developer - I didn’t even know you could do git patches via email until recently. And while it’s super cool, it also sounds kinda terrible, especially at the volume they must be receiving? Their own docs are saying the mailing lists receive some 500 emails per day and I can’t imagine the merge process is fun.
So many doc pages are dedicated to how to submit a patch - which is great that it’s documented, and I’m sure it will always be somewhat complicated for a large project - but it also feels like things that are all automatically handled by newer tools / bots which can automatically enforce style checks, etc.
I guess they could argue that the complicated process acts as a filter to people submitting PRs who don’t know what they are doing, but I’d argue it also shuts out talented engineers who don’t have 40 hours to learn how to submit a patch to a project on top of also learning the kernel and also fixing the bug in question.
From what little I read of their git process, does anyone know if there’s anything preventing the maintainer of a subsystem from setting up a more modern method for receiving patches? As long as the upstream artifact to the kernel has the expected format?
Oh man, I actually like the language, but you made me think of my own hot take:
Python has inexcusably poor docs.
Just a smattering of examples, which aren’t even that good, while failing to report key information like all the parameters a function can take, or all the exceptions it can throw. Any other popular language I can think of has this locked down and it makes things so much easier.
I made one such tool!
https://github.com/CMahaff/lasim
I know there’s also a python script out there and a new Android app that has active syncing. I don’t have links handy to those on mobile.
Excellent work by the way, much better solution than my own :)
Curious did you get the survey popup in desktop mode on the deck? Or does it work in “big picture”?