Post

I Made Friends With My Local E-Waste Guy And Mentioned That I Wanted To Start A Homelab

I Made Friends With My Local E-Waste Guy And Mentioned That I Wanted To Start A Homelab

Introduction

The blinking lights of decommissioned enterprise hardware have a peculiar allure for infrastructure engineers. When I casually mentioned my homelab ambitions to my local e-waste technician, I didn’t expect to walk away with a treasure trove of Dell PowerEdge servers, Cisco switches, and vintage Sun SPARCstations. This serendipitous haul presented both an incredible opportunity and a formidable challenge: how to transform enterprise cast-offs into a fully functional DevOps training environment.

Homelabs serve as critical infrastructure for skill development in system administration and DevOps engineering. Unlike cloud-based solutions, physical hardware provides direct exposure to:

  • Bare-metal provisioning
  • Enterprise-grade networking
  • Hardware failure scenarios
  • Power and thermal management

In this comprehensive guide, we’ll transform e-waste into enterprise-grade infrastructure covering:

  1. Hardware assessment and component verification
  2. Enterprise network configuration at scale
  3. Hypervisor deployment on heterogeneous hardware
  4. Sustainable power management
  5. Legacy system integration strategies

For DevOps professionals, hands-on hardware experience provides irreplaceable insights into:

  • Infrastructure-as-Code limitations in physical environments
  • Real-world performance characteristics of storage subsystems
  • Network latency implications in distributed systems
  • Hardware-level security considerations

Understanding Homelab Infrastructure Fundamentals

What Constitutes an Enterprise Homelab?

A professional-grade homelab differs from typical home servers through its implementation of:

  • Enterprise Reliability Features: ECC memory, hardware RAID, redundant PSUs
  • Network Segmentation: VLANs, firewall zones, management interfaces
  • Centralized Management: IPMI/iDRAC, KVM over IP, serial consoles
  • Service Grading: Production vs. experimental workload isolation

Historical Context of Decommissioned Hardware

The equipment typically available through e-waste channels follows a predictable lifecycle:

  1. 3-5 years: Production duty in enterprise environments
  2. 2-3 years: Cold standby or DR use
  3. 1-2 years: Decommissioned but warehoused
  4. E-Waste Phase: Sold to recyclers at ~10-15% original value

Our acquired equipment falls into this final phase:

  • Dell PowerEdge Generation:
    • 11th Gen (R710): 2010 Nehalem/Westmere architecture
    • 12th Gen (R620): 2012 Ivy Bridge
    • 13th Gen (R530/T420): 2014 Haswell
  • Cisco 3850 Series: Current generation Catalyst switches with UPOE
  • Legacy Systems: SPARCstations and PDP-11 serve as historical artifacts rather than production systems

Technical Evaluation Criteria

Each component requires rigorous assessment before deployment:

Servers:

1
2
3
4
# Check critical hardware components
sudo lshw -class memory -class storage -class power -class processor
sudo smartctl -a /dev/sda # Repeat for all drives
ipmitool sel list # View system event log

Switches:

1
2
3
show version # Check IOS version and uptime
show environment all # Verify cooling and power
show interfaces status # Validate physical ports

Power and Thermal Considerations

Enterprise equipment prioritizes performance over efficiency:

  • R620 Power Consumption:
    • Idle: 120-150W
    • Load: 300-400W
  • Heat Output:
    • 1U server ≈ 1000 BTU/hr at load
    • Rack density planning: 20-30kW per cabinet in enterprise vs. 2-5kW for homelabs

Prerequisites for Enterprise Homelab Deployment

Hardware Requirements

Our e-waste inventory forms a complete stack:

ComponentQuantityRole
R5302Virtualization hosts (Proxmox)
R6201Storage server (TrueNAS)
T4201Backup target
Cisco 3850-48-UPOE7Core/distribution switching
Liebert GXT31Power conditioning

Network Architecture

Professional-grade segmentation is essential:

VLAN Architecture:

1
2
3
4
5
VLAN 10: Management (IPMI/iDRAC)
VLAN 20: Hypervisor Communication
VLAN 30: VM Infrastructure
VLAN 40: Storage Network (iSCSI/NFS)
VLAN 50: Guest/IoT Isolation

Switch Port Configuration:

1
2
3
4
5
interface GigabitEthernet1/0/1
 description Proxmox-01 Management
 switchport access vlan 10
 switchport mode access
 power inline never # Disable PoE on server ports

Pre-Installation Checklist

  1. Hardware Validation:
    • Memtest86+ for 24 hours minimum
    • Drive burn-in with badblocks:
      1
      
      badblocks -wsv /dev/sda
      
  2. Firmware Updates:
    1
    2
    3
    4
    
    # Dell Update Packages
    sudo ./SUU_LIN64.bin --update
    # Cisco IOS Upgrade
    copy tftp: flash:c3850-universalk9-tar.xxxxxx.tar
    
  3. Physical Infrastructure:
    • 20A dedicated circuits (NEMA 5-20R)
    • Sound-dampened rack enclosure
    • Structured cabling (Cat6A for 10GBase-T)

Installation and Configuration Walkthrough

Hypervisor Deployment on R530

Proxmox VE Installation:

1
2
3
4
5
6
7
# Prepare RAID array
sudo perccli /c0 set jbod=off
sudo perccli /c0 add vd type=raid1 drives=0:0:0,0:0:1 size=all

# Install Proxmox
sudo dpkg-reconfigure locales
sudo apt-get install proxmox-ve postfix open-iscsi

Network Configuration:

1
2
3
4
5
6
7
8
9
10
11
# /etc/network/interfaces
auto eno1
iface eno1 inet manual # Management interface

auto vmbr0
iface vmbr0 inet static
  address 192.168.10.2/24
  gateway 192.168.10.1
  bridge-ports eno2
  bridge-stp off
  bridge-fd 0

Cisco 3850 Switch Stack Configuration

Base Stack Setup:

1
2
3
4
5
# Initialize switch stacking
switch 1 provision ws-c3850-48upoe
switch 2 provision ws-c3850-48upoe
switch 1 priority 15 # Make primary
reload slot 1 # Rebuild stack

Critical Security Settings:

1
2
3
4
5
enable secret 9 $STRONG_PASSWORD
no ip http server
no ip http secure-server
service password-encryption
logging buffered 16384 informational

UPS Integration with NUT

Network UPS Tools Configuration:

1
2
3
4
5
# /etc/nut/ups.conf
[liebert]
    driver = usbhid-ups
    port = auto
    desc = "Liebert GXT3 700VA"

Cluster-Aware Monitoring:

1
2
3
# /etc/nut/upsmon.conf
MONITOR liebert@localhost 1 monuser $PASSWORD master
SHUTDOWNCMD "/sbin/shutdown -h +0"

Infrastructure Optimization Techniques

Performance Tuning for Older Hardware

BIOS Settings (Dell PowerEdge):

  1. Power Management: Static High Performance
  2. Turbo Boost: Disabled (reduces thermal load)
  3. C-States: C1 only
  4. Memory Speed: Maximum supported (1866MHz for DDR3)

Linux Kernel Tuning:

1
2
3
4
5
6
# /etc/sysctl.d/99-homelab.conf
vm.swappiness=10
vm.dirty_ratio=40
vm.dirty_background_ratio=10
net.core.rmem_max=16777216
net.core.wmem_max=16777216

Enterprise Storage Configuration

ZFS Pool Design for R620:

1
2
3
4
5
# Create mirrored vdev pool
zpool create tank mirror /dev/disk/by-id/ata-XXXXXX /dev/disk/by-id/ata-YYYYYY
zfs set compression=lz4 tank
zfs set atime=off tank
zfs set recordsize=1M tank/vmstorage

Network Security Hardening

Cisco ACL Implementation:

1
2
3
4
5
6
# Restrict management access
ip access-list extended MGMT_ACL
 permit tcp 192.168.10.0 0.0.0.255 any eq 22
 deny ip any any log
interface Vlan10
 ip access-group MGMT_ACL in

Hypervisor Firewall Rules:

1
2
3
# Proxmox host firewall
pve-firewall compile # Generate rules
pve-firewall update # Apply changes

Operational Management Procedures

Infrastructure Monitoring Stack

Prometheus Node Exporter Setup:

1
2
3
# Install on Proxmox hosts
sudo apt install prometheus-node-exporter
systemctl enable --now prometheus-node-exporter

Grafana Dashboard Configuration:

1
2
3
4
5
# Homelab dashboard variables
variables:
  - name: host
    query: label_values(node_uname_info{job="node"}, instance)
    refresh: 2

Automated Backup Strategy

Proxmox Backup Configuration:

1
2
3
# vzdump hook script
vzdump 100 --mode snapshot --storage nas-backups \
  --exclude-path /mnt/temporary_data/ --mailto admin@domain.tld

Versioned Backups with Borg:

1
2
3
# Daily backup cron job
borg create --stats --progress /backup/hostname::'{now:%Y-%m-%d}' \
  /etc /var/lib/important-data

Troubleshooting Enterprise Homelabs

Common Failure Scenarios

Problem: High CPU Ready Times in VMs
Diagnosis:

1
2
3
# Check CPU contention
qm config $VMID | grep cores
mpstat -P ALL 1 5 # Check host CPU usage

Solution: CPU pinning or core allocation adjustments

Problem: Switch Stack Flapping
Diagnosis:

1
2
show switch # Verify stack membership
show interface transceiver # Check SFP health

Solution: Replace faulty stack cables or upgrade IOS

Hardware Diagnostics

Memory Errors:

1
2
# Dell-specific diagnostics
sudo /opt/dell/srvadmin/bin/omreport chassis memory

Drive Failures:

1
2
3
4
# SMART short test
sudo smartctl -t short /dev/sda
# Check results
sudo smartctl -l selftest /dev/sda

Conclusion

Transforming e-waste into enterprise-grade infrastructure provides unparalleled learning opportunities that cloud environments simply can’t replicate. Through this process, we’ve:

  1. Established a production-like environment with proper network segmentation
  2. Implemented enterprise management practices (IPMI, centralized logging)
  3. Created a sustainable power management solution
  4. Demonstrated real-world hardware troubleshooting techniques

This homelab serves as an ideal platform for exploring advanced DevOps concepts:

For further learning, consult these essential resources:

The true value of a homelab lies not in the hardware itself, but in the operational knowledge gained through hands-on experience with enterprise-grade systems - experience that directly translates to professional DevOps and infrastructure management roles.

This post is licensed under CC BY 4.0 by the author.