• 1 Post
  • 19 Comments
Joined 1 year ago
cake
Cake day: June 11th, 2023

help-circle



  • Honestly I just picked it here because I thought it’d look nice giving a big hit of green color and gives a good texture addition since everything else is pretty soft, this adds some crunch to the dish. For me there’s no vegetable that really stands out to me where I will look forward to it, except maybe roast potatoes lol




  • It sounds like what you need to do at this point is find what IP address your lemmy instance and mastodon instance containers are using on your VPS. you can do “docker inspect containername” and look for the IP address in there. it might be something like 172.16.0.1 for lemmy and 172.17.0.1 for mastodon. then you want to set up your reverse proxy to point lmy.my-domain.tld to 172.16.0.1:80 (or whatever port you set lemmy to use) and then mstdn.my-domain.tld to point to 172.17.0.1:80 (again, port might be different, i dont know what the default port is)

    -IF- both of the containers are using the same IP, then you will need to make sure that they are using different ports. if they are on the same ip and same port, whichever container loads 2nd will fail to properly load, because when a port is taken on an IP address, it is reserved and nothing else can try to listen on that port.


  • So a reverse proxy is sort of like a phonebook or directory, it routes outside requests to the appropriate place. So imagine your reverse proxy is a receptionist, someone comes in and says “hey I am looking for plex.mydomain.com” the receptionist would then use the phonebook and say “ok if you are looking for plex.mydomain.com, go to building 192.168.1.10 (the ip), room 9000 (the port)”

    Since you are asking about dockerized services, the networking for those can be done in several different ways, but the one thing that really matters is that each service needs to have a unique combination of ip and port, because only 1 service can live at each address. With docker, you could set up multiple services that use the host server’s ip, in which case each container will need to be on different ports, or you could have it so each container has its own ip, in which case the port can be anything.






  • MongoDB has a modified version of the AGPL that they call Server Side Public License that might interest you. Specifically the change in section 13:

    “Service Source Code” means the Corresponding Source for the Program or the modified version, and the Corresponding Source for all programs that you use to make the Program or modified version available as a service, including, without limitation, management software, user interfaces, application program interfaces, automation software, monitoring software, backup software, storage software and hosting software, all such that a user could run an instance of the service using the Service Source Code you make available.

    By my reading, it closes that loophole you mention by specifically calling out interfaces and APIs as also requiring the source to be available. At the top of the page I linked there is also a PDF showing the removals and additions they made to the AGPL to end up with their SSPL.




  • kbin and lemmy are different softwares, but they are both used for link aggregation and the 2 softwares use a common protocol, so they can talk to each other. So there are kbin servers and lemmy servers, and they are all interconnected.

    So now we can take this post as an example:

    You are a user on kbin.social

    You posted this question on /m/nostupidquestions@lemmy.world - this means the community you posted on is actually hosted on lemmy.world

    lemmy.world then tells other instances that its federated with that someone just made a post on /c/nostupidquestions on its instance. what kbin calls magazines are called communities on lemmy, hence the /c/ instead of the /m/.

    kbin.social and all the other instances will then also show this post, even though it originally was created on a different instance




  • Typically you’d have a server running on a device on each UPS, and the clients would be the other devices also plugged into that UPS, so when that UPS is low, everything plugged into it will turn off. If you have another UPS elsewhere in the house, you would have another server installed on a device there so it can monitor that UPS that it is plugged into, and tell the other devices also plugged into that other UPS to shut down. Without knowing the layout you are running though, there’s no way to know if NUT is what you need or want.

    So in your case it would likely be to plug your server and nas both into the same UPS, and when the server detects the UPS is low battery, it will tell the NAS to shut down. This would also require the switches/router/whatever to also be on a UPS to hold power of course. So then it basically becomes each little cluster of devices that need UPS would also have a switch nearby that is also on the UPS