gargravarr2112

joined 11 months ago
[–] gargravarr2112@alien.top 1 points 10 months ago

Can recommend my APC SMT1500I. Original batteries lasted 9.5 years. Accidentally plugged a fan heater into it once and it survived. Most reliable device in my rack.

[–] gargravarr2112@alien.top 1 points 10 months ago

If someone or something malicious gets a shell account on my systems, then it at least stops them doing anything system-wide. And yes, if a script is going to request admin rights to do something, it'll stop right at the sudo prompt. Passwordless, it could do stuff without you even being aware of it.

Whether or not this is a line of defence at all is open to debate.

[–] gargravarr2112@alien.top 1 points 10 months ago

Heavy computation rack is in an unheated conservatory with a window cracked open. Keeps the HDD temperatures around 30 degrees. Temperature monitoring from my PDU shows a 3'C rise from the inlet to the exhaust side of the rack. This stuff is mostly powered off when not in use. In summer, it can get to 35'C in that room so I shut everything down at that point.

24/7 rack is in my lounge and vents the heat into the room (helps a little bit with heating costs). Top of the rack is about 37'C but I've seen it around 45'C with all my hypervisors doing stuff. Nothing complains. As long as the intake air is within the manufacturer's stated range, it's fine.

Might want to consider redirecting the heat into the house rather than venting it outside.

[–] gargravarr2112@alien.top 1 points 10 months ago (2 children)
  1. Domain auth (1 place to set passwords and SSH keys), no root SSH
  2. SSH by key only
  3. Passworded sudo (last line of defence)
  4. Only open firewall hole is OpenVPN with security dialled up high
  5. VLANs - laptops segregated from servers
  6. Strict firewall rules between VLANs
  7. TLS on everything
  8. Daily update check alerts (no automatic updates, but persists until I deal with them)
  9. Separate isolated syslog server for audit trails
  10. Cold backups
[–] gargravarr2112@alien.top 1 points 10 months ago (1 children)

DIY - No Regrets.

I built my NAS out of spare parts originally and then it evolved into needing dedicated purchases. I like having full control of the OS and everything on it - it helps me understand what daemons are doing what. It does a lot more than file sharing.

The likes of QNAP and Synology may make a more polished product with an easy UI, as well as offering support, but as far as I care, I am support, so I like to fix problems myself.

If you're ping-ponging between the two options, from your post it reads like cost is the biggest problem you face. But as you say, storage is a critical part of the infrastructure and sometimes you do have to spend money on it if you want it to be reliable. I just upgraded my main NAS with a larger chassis and motherboard (from an ITX) so I can expand it further. It cost me a sizeable amount of money that might have bought me a low-end ready-made, but this is far more flexible.

[–] gargravarr2112@alien.top 1 points 10 months ago (1 children)

Yes, this should work fine. SAS does not care what path the signal takes - it doesn't differentiate between internal and external. You can run internal over external cables without issue. I've done similar by turning my old NAS chassis into a DAS, and connecting it to her internal ports of the HBA. And you can connect SAS or SATA drives to the DAS (system 1).

[–] gargravarr2112@alien.top 1 points 10 months ago (1 children)

Why not just use ThinOS? IIRC it supports the NX protocol that NoMachine uses. If not, then you have Wyse's respin of Ubuntu called ThinLinux, which you should be able to install.

You can download the OS image and flash utility from Dell's website.

[–] gargravarr2112@alien.top 1 points 10 months ago

Can you tell if the backplane supports SGPIO? If it and your RAID card do, and things are wired up correctly, you can tell your RAID card to Identify the drive and it'll blink the LEDs. I just configured a Supermicro chassis to do this with an Adaptec controller.

[–] gargravarr2112@alien.top 1 points 10 months ago (1 children)

Several locations.

  • bunch of metering smart plugs flashed with Tasmota, feeding into Home Assistant (power at the wall)
  • UPS being polled by LibreNMS (difference in load Vs wall shows the UPS consumption)
  • metering PDU polled by LibreNMS shows the server load
  • PMBus PSUs in the servers report their own power consumption to the BMC

And after all this, I do... Nothing with this data.

[–] gargravarr2112@alien.top 1 points 10 months ago (1 children)

Power in the UK has gone through the roof. I've downsized my lab as much as I can and have at times wondered if I should shut it down completely.

Originally I was running an EdgeRouter 4, Zyxel 48-port managed switch and custom-built NAS with an i3-9100T, 32GB ECC and 6x 12TB SAS drives in a zpool. The NAS did everything - VMs, storage, backups etc. but it was pulling quite a lot of power.

A while back I ran a USFF PC as my server, which idled at 8W. Versus my 200W Xeon machine at the time, it paid for itself in 12 months. I dug that out and moved the VMs onto it. Storage went onto an ARM NAS. I was running too many VMs for a single USFF even maxed out, so I bought another 2 of them (identical). Now I run them in a Proxmox cluster. I use a passive cooled HP 1810 managed switch and an EdgeRouter Lite for the network, plus an Apple Airport with its transmitter dialled down to 25%. The ARM machine is much slower than my ZFS NAS, but it is much lighter on power - at that point, the HDDs are the significant draw, so I only run 2 spinners that are non-redundant and make sure they're backed up to cold storage. I also power up my ZFS machine once a month or so and sync the data from it. Other than that, I keep the big x86 machines shut down until needed.