TCB13

joined 2 years ago
[–] [email protected] 3 points 4 days ago

The problem with compiled JS is that it doesn't last long. Any piece of software written in the popular frameworks will probably not compile in 5 years because xyz dependency. There might be ways around it, but all annoying and not good. Some type of software really needs more assurances that "might compile in 5 years" because some people can't afford to upgrade to the latest framework down the path due to size or simply lack of time.

[–] [email protected] 2 points 4 days ago* (last edited 4 days ago) (2 children)

📦 Zero Dependencies - Pure JavaScript, no build step required

This is really the best part. However the example does run a npm run build...

I would be great to have something that can be imported in a script tag like jquery and not something that requires npm and compilarion.

[–] [email protected] -4 points 4 days ago

I don't really get this type of "media" bullshit articles. Yes, Windows is becoming progressively worse with more annoyances but you also have more simple to use tools than ever to disable those annoyances in bulk.

For the average user is far simpler to just run W10 Privacy, CTT or some other tool to disable all the annoying Windows features than it is to move to Linux and face all the major pain points people usually have around software compatibility and missing xyz very specific that isn't really the same thing under Linux.

There you go, fixed the Windows problem for you in a few clicks, no need to download an entire new OS and complain afterwards.

[–] [email protected] 2 points 5 days ago* (last edited 5 days ago)

If you want to get the job (of publishing a blog) done fast, and move on, then use WP. If you want to mess around probably look at editorjs.io.

[–] [email protected] 2 points 5 days ago

If you want to get the job (of publishing a blog) done fast, and move on, then use WP. If you want to mess around probably look at editorjs.io.

[–] [email protected] 4 points 5 days ago

At least WP is free, Ghost is as "free" until you find out its only useful with the rest of the payed platform. editorjs.io is much better in that sense.

[–] [email protected] 4 points 5 days ago* (last edited 5 days ago)

At least archive.today actually works to bypass paywalls... and provides content quickly. archive.org is massive and cool but it usually doesn't contain snapshots from paid articles, it is also very very slow, US-controlled and the way you look for a snapshot and move the dates is painfully slow.

[–] [email protected] 1 points 1 week ago

Yeah Microsoft for what's worth does play ball, you can open complaints and they'll actually read those and act fast. Google is a total pain to deal with, even if you're on some type of google partnership they'll not do much.

[–] [email protected] 1 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

I don't disagree with you but... it also provides a cohesive ecosystem of tools to manage linux. What we had before was a poorly integrated mess of smaller tools that was just too hard to maintain and sometimes use.

Besides not all systemd components come out of the box with the base binary, some have to be installed if you need them. And no, it doesn't get in the way. :)

 

cross-posted from: https://lemmy.world/post/29924695

If compatibility, performance, and long-term security are important to you, Windows 10 Enterprise + ESU is the more future-proof route to go - all without sacrificing the "clean" LTSC experience.

 

If compatibility, performance, and long-term security are important to you, Windows 10 Enterprise + ESU is the more future-proof route to go - all without sacrificing the "clean" LTSC experience.

 

You may hate me for this, but this is what’s really going on. I love JS/TS and Node but the world is built in economies of scale not in love for programming languages… and PHP is the best when it comes to cheap scaling.

Really hope someone at Node decides to make it fast-cgi compatible in a nice way.

 

You may hate me for this, but this is what's really going on. I love JS/TS and Node but the world is built in economies of scale not in love for programming languages... and PHP is the best when it comes to cheap scaling.

Really hope someone at Node decides to make it fast-cgi compatible in a nice way.

 

cross-posted from: https://lemmy.world/post/28692919

uSentry is a lightweight, self-hosted Identity and Access Management (IAM) and Single Sign-On (SSO) solution designed for homelab and small-scale environments.

⚡ A single PHP file. < 400 lines of code. No database. No background processes. No cloud. Just works. ⚡

Most IAM and SSO solutions require databases, certificates and background services baked into a dozen containers. This is all fine but also also overkill for homelabs and impossible for low-power ARM devices. uSentry is different, it isn't pretty but it sucks less for a lot of use cases.

Enjoy!

 

cross-posted from: https://lemmy.world/post/28692919

uSentry is a lightweight, self-hosted Identity and Access Management (IAM) and Single Sign-On (SSO) solution designed for homelab and small-scale environments.

⚡ A single PHP file. < 400 lines of code. No database. No background processes. No cloud. Just works. ⚡

Most IAM and SSO solutions require databases, certificates and background services baked into a dozen containers. This is all fine but also also overkill for homelabs and impossible for low-power ARM devices. uSentry is different, it isn't pretty but it sucks less for a lot of use cases.

Enjoy!

 

cross-posted from: https://lemmy.world/post/28692919

uSentry is a lightweight, self-hosted Identity and Access Management (IAM) and Single Sign-On (SSO) solution designed for homelab and small-scale environments.

⚡ A single PHP file. < 400 lines of code. No database. No background processes. No cloud. Just works. ⚡

Most IAM and SSO solutions require databases, certificates and background services baked into a dozen containers. This is all fine but also also overkill for homelabs and impossible for low-power ARM devices. uSentry is different, it isn't pretty but it sucks less for a lot of use cases.

Enjoy!

 

uSentry is a lightweight, self-hosted Identity and Access Management (IAM) and Single Sign-On (SSO) solution designed for homelab and small-scale environments.

⚡ A single PHP file. < 400 lines of code. No database. No background processes. No cloud. Just works. ⚡

Most IAM and SSO solutions require databases, certificates and background services baked into a dozen containers. This is all fine but also also overkill for homelabs and impossible for low-power ARM devices. uSentry is different, it isn't pretty but it sucks less for a lot of use cases.

Enjoy!

 

cross-posted from: https://lemmy.world/post/23071801

Considering a lot of people here are self-hosting both private stuff, like a NAS and also some other is public like websites and whatnot, how do you approach segmentation in the context of virtual machines versus dedicated machines?

This is generally how I see the community action on this:

Scenario 1: Fully Isolated Machine for Public Stuff

Two servers one for the internal stuff (NAS) and another for the public stuff totally isolated from your LAN (websites, email etc). Preferably with a public IP that is not the same as your LAN and the traffic to that machines doesn't go through your main router. Eg. a switch between the ISP ONT and your router that also has a cable connected for the isolated machine. This way the machine is completely isolated from your network and not dependent on it.

Scenario 2: Single server with VM exposed

A single server hosting two VMs, one to host a NAS along with a few internal services running in containers, and another to host publicly exposed websites. Each website could have its own container inside the VM for added isolation, with a reverse proxy container managing traffic.

For networking, I typically see two main options:

  • Option A: Completely isolate the "public-facing" VM from the internal network by using a dedicated NIC in passthrough mode for the VM;
  • Option B: Use a switch to deliver two VLANs to the host—one for the internal network and one for public internet access. In this scenario, the host would have two VLAN-tagged interfaces (e.g., eth0.X) and bridge one of them with the "public" VM’s network interface. Here’s a diagram for reference: https://ibb.co/PTkQVBF

In the second option, a firewall would run inside the "public" VM to drop all inbound except for http traffic. The host would simply act as a bridge and would not participate in the network in any way.

Scenario 3: Exposed VM on a Windows/Linux Desktop Host

Windows/Linux desktop machine that runs KVM/VirtualBox/VMware to host a VM that is directly exposed to the internet with its own public IP assigned by the ISP. In this setup, a dedicated NIC would be passed through to the VM for isolation.

The host OS would be used as a personal desktop and contain sensitive information.

Scenario 4: Dual-Boot Between Desktop and Server

A dual-boot setup where the user switches between a OS for daily usage and another for hosting stuff when needed (with a public IP assigned by the ISP). The machine would have a single Ethernet interface and the user would manually switch network cables between: a) the router (NAT/internal network) when running the "personal" OS and b) a direct connection to the switch (and ISP) when running the "public/hosting" OS.

For increased security, each OS would be installed on a separate NVMe drive, and the "personal" one would use TPM with full disk encryption to protect sensitive data. If the "public/hosting" system were compromised.

The theory here is that, if properly done, the TPM doesn't release the keys to decrypt the "personal" disk OS when the user is booted into the "public/hosting" OS.

People also seem to combine both scenarios with Cloudflare tunnels or reverse proxies on cheap VPS.


What's your approach / paranoia level :D

Do you think using separate physical machines is really the only sensible way to go? How likely do you think VM escape attacks and VLAN hopping or other networking-based attacks are?

Let's discuss how secure these setups are, what pitfalls one should watch out for on each one, and what considerations need to be addressed.

 

Considering a lot of people here are self-hosting both private stuff, like a NAS and also some other is public like websites and whatnot, how do you approach segmentation in the context of virtual machines versus dedicated machines?

This is generally how I see the community action on this:

Scenario 1: Air-gapped, fully Isolated Machine for Public Stuff

Two servers one for the internal stuff (NAS) and another for the public stuff totally isolated from your LAN (websites, email etc). Preferably with a public IP that is not the same as your LAN and the traffic to that machines doesn't go through your main router. Eg. a switch between the ISP ONT and your router that also has a cable connected for the isolated machine. This way the machine is completely isolated from your network and not dependent on it.

Scenario 2: Single server with VM exposed

A single server hosting two VMs, one to host a NAS along with a few internal services running in containers, and another to host publicly exposed websites. Each website could have its own container inside the VM for added isolation, with a reverse proxy container managing traffic.

For networking, I typically see two main options:

  • Option A: Completely isolate the "public-facing" VM from the internal network by using a dedicated NIC in passthrough mode for the VM;
  • Option B: Use a switch to deliver two VLANs to the host—one for the internal network and one for public internet access. In this scenario, the host would have two VLAN-tagged interfaces (e.g., eth0.X) and bridge one of them with the "public" VM’s network interface. Here’s a diagram for reference: https://ibb.co/PTkQVBF

In the second option, a firewall would run inside the "public" VM to drop all inbound except for http traffic. The host would simply act as a bridge and would not participate in the network in any way.

Scenario 3: Exposed VM on a Windows/Linux Desktop Host

Windows/Linux desktop machine that runs KVM/VirtualBox/VMware to host a VM that is directly exposed to the internet with its own public IP assigned by the ISP. In this setup, a dedicated NIC would be passed through to the VM for isolation.

The host OS would be used as a personal desktop and contain sensitive information.

Scenario 4: Dual-Boot Between Desktop and Server

A dual-boot setup where the user switches between a OS for daily usage and another for hosting stuff when needed (with a public IP assigned by the ISP). The machine would have a single Ethernet interface and the user would manually switch network cables between: a) the router (NAT/internal network) when running the "personal" OS and b) a direct connection to the switch (and ISP) when running the "public/hosting" OS.

For increased security, each OS would be installed on a separate NVMe drive, and the "personal" one would use TPM with full disk encryption to protect sensitive data. If the "public/hosting" system were compromised.

The theory here is that, if properly done, the TPM doesn't release the keys to decrypt the "personal" disk OS when the user is booted into the "public/hosting" OS.

People also seem to combine both scenarios with Cloudflare tunnels or reverse proxies on cheap VPS.


What's your approach / paranoia level :D

Do you think using separate physical machines is really the only sensible way to go? How likely do you think VM escape attacks and VLAN hopping or other networking-based attacks are?

Let's discuss how secure these setups are, what pitfalls one should watch out for on each one, and what considerations need to be addressed.

 

cross-posted from: https://lemmy.world/post/21563379

Hello,

I'm looking for a high resolution image of the PAL cover from the Dreamcast (I believe).

There was this website covergalaxy that used it have in 2382x2382 but all the content seems to be gone. Here's the cache https://ibb.co/nRMhjgw . Internet archive doesn't have it.

Much appreciated!

 

Hello,

I'm looking for a high resolution image of the PAL cover from the Dreamcast (I believe).

There was this website covergalaxy that used it have in 2382x2382 but all the content seems to be gone. Here's the cache https://ibb.co/nRMhjgw . Internet archive doesn't have it.

Much appreciated!

view more: next ›