skilltheamps

joined 2 years ago
[–] [email protected] 6 points 9 months ago (1 children)

"almost all of the most technical employees in framework are using either ubuntu, fedora or nixos. I'm mostly on Windows because we need actually people that are using Windows because our employee base in framework is all Linux users"

  • Nirav Patel

https://m.youtube.com/watch?v=EIEc43CxIvY

[–] [email protected] 10 points 9 months ago (7 children)

That is not the case for every country though. In France and Germany for example almost 3/4 of google requests are via IPv6.

[–] [email protected] 1 points 1 year ago

Well you must have either set up a port redirect (ipv4) or opened the port for external traffic (ipv6) yourself. It is not reachable by default as home routers put a NAT between the internet and your devices, or in the case of ipv6 they block any requests. So (unless you have a very exotic and unsafe router) just uhhh don't 😅 To serve websites it is enough to open 443 for https, and possibly 80 for http if you want to serve an automatic redirect to https.

[–] [email protected] 1 points 1 year ago (1 children)

That's odd, I upgraded my ender 3 with bed leveling and removed the knobs to mount it fixed, because the damn knobs keep moving and then you have to redo the bed calibration. To be honest I can imagine one reason might be that a loosely mounted bed gives you more fault tolerance against the nozzle being too low. I put my bed on two parallel linear rollers for more rigidity, and combined with dual z screws the nozzle has no chance anymore to produce any sort of first layer when it is slightly too low. That made me realize just how much the stock ender 3 is flopping around, but also how this can give you mostly okayish results most of the time without having to deal with a ton of small tolerances.

[–] [email protected] 16 points 1 year ago (2 children)

A colleague of mine had a (non externally reachable) raspberry pi with default credentials being hijacked for a botnet by a infected windows computer in the home network. I guess you'll always have people come over with their devices you do not know the security condition of. So I've started to consider the home network insecure too, and one of the things I want to set up is an internal ssh honeypot with notifications, so that I get informed about devices trying to hijack others. So for this purpose that tool seems a possibilty, hopefully it is possible to set up some monitoring and notification via uptime kuma.

[–] [email protected] 5 points 1 year ago* (last edited 1 year ago) (1 children)

Well it is compiled to byte code in a first step, and this byte code then gets processed by the interpreter. Now Java does the exact same thing: gets compiled to byte code which then gets executed by the jvm (java virtual machine), which is essentially a interpreter that is just a little simpler than the python one (has fewer types for example). And yet, nobody talks about a java interpreter

[–] [email protected] 27 points 1 year ago (1 children)

You do not want Octoprint on a machine that is busy. Otherwise you have load spikes that cause Octoprint to not be able to send the move-commands (gcode) as fast as the printer executes the movements. This problem is pronounced with faster printers and slicers that break up arcs into small straight lines (which is practically all slicers). Otherwise your printer stutters because it has to take small breaks to wait for the next command from octoprint.

[–] [email protected] 5 points 1 year ago (3 children)

What privacy concerns do you have? I'm all for privacy, but I don't really see where registrars are a delicate topic in that. The most that comes to mind is that some (most?) have a service where they do not give out your name and address for whois requests, but instead the details of the registrar (namecheap has that for example).

[–] [email protected] 2 points 1 year ago

True words. The sustained effort to keep something in decent shape over years is not to be underestimated. Now when life changes and one is not able or willing anymore to invest that amount of time, ill-timed issues can become quite the burden. At one point I decided to cut down on that by doing a better founded setup, that does backup with easy rollback automatically, and updates semi-automatically. I rely on my server(s), and all from having this idea to having it decently implemented took me a number of months. Just because time for such activities is limited, and getting a complex and intertwined system like this reliably and fault tolerant automated and monitored is simply something else than spinning up a one off service

[–] [email protected] 12 points 1 year ago (1 children)

And they believe all employees actually remember so many wildly different and long passwords, and change them regularly to wildly different ones? All this leads to is a single password that barely makes it over the minimum requirements, and a suffix for the stage (like 1 for boot, 2 for bitlocker etc), and then another suffix for the month they changed it. All of that then on sticky notes on the screen.

[–] [email protected] 4 points 1 year ago

Since you run everything in docker, I guess you have experienced the benefits of containerization. So why not leverage that for your host too?

Fedora IoT is a container-based host that runs on your hardware, with a focus on edge device deployment.

https://fedoraproject.org/iot/ I have it running on two servers as well, and it works great. The only thing I changed is that I layered docker on it instead of using podman, because at the time I had trouble getting my reverse proxy working properly over ipv6

[–] [email protected] 10 points 1 year ago (1 children)

Not to me. Absence of QA allows faulty parts to make it into a plane, it does not explain why there are faults in the first place. For doors and wheels popping off there have to be either lethal part design mistakes, parts made from play doh instead of aluminium/steel, or the people on the assembly line throwing fasteners in the bin instead of putting them on. It's not like a door pops of because its seal touched soap once and somebody poked an unverified piece of plastic at it. Especially in aviation, where you need to have redundancies.

 

Hello,

I moved my home servers to fedora silverblue and docker-compose (ipv6 reasons :/). I stumpled upon the problem that I neither wanted to update image tags manually, nor have no idea what ":latest" deployed on my server in case I need to roll back.

To alleviate that problem, I made a small update-tool. It takes care of writing down the image@sha256... digest every time so that you can roll back. It also automatically snapshots and restarts the services.

It is made in Python but doesn't need any dependencies, so no catering for a venv either. You only need to have skopeo and snapper in working order. Maybe you'll find it useful, but please be aware that it is in an early stage. Also I'm not responsible if it nukes your server 😅

 

I often observe that people that started a small open source project seem to abandon it sooner or later. I'm guilty of this myself in numerous cases. Reasons there are many probably, from new obligations in life to shifts in interest and whatnot.

At some point somebody comes by with an issue, or a merge request even, but the maintainer does not take care of it. Usually this ends up in forks, often though forks undergo the same fate. Apart from the immediate forks-jungle, stuff like software stores or other things might be hardlinked to the original repo, which means places like these end up with dead originals and a number of forks with varying degree of being maintained as well.

To me its just a sad situation overall. And yet I cannot find the time or motivation to maintain some stuff, because circumstances just changed. And I also do not think one is obliged to do so, just because they where nice enough to share their code when the project mattered to them.

Is there a better way? Usually these are very nieche projects, and there is not a circle of regularly active developers that could share administration of a repo, but rather a quiet one-man-show with a short timespan of incredible activity. Some kind of sensible failover mechanism once the original maintainer vanishes would probably be cool. Or any other way that introduces some redundancy in keeping a repository alive. You know how package maintainers in Linux distributions open their package(s) for adoption by somebody else if they run out of capacity? I think that is nice.

I will publish a small project soon I think, but somewhere in the future I fear to leave one or the other person frustrated again when I have moved on to other things...

view more: next ›