AssuredSAN/Quantum QXS iSCI/FiberChannel SAN/NAS

I had the chance to re-purpose a Quantum QXS 1200 2U12 4th Generation Storage array.  I am just learning here, so take everything I say here with a grain of salt.  This post will continue to develope as I learn and test some more.

These are actually manufactured by DotHill Corporation, which Seagate acquired.  Quantum has some special firmware on the machines, and targeted them to fit into their StorNext platform.

The one I am working on as previously mentioned is a QXS 2U12, which is equilvant to a AssuredSAN 4xxx series.  It has twelve 3TB 3.5" 7200rpm SAS drives in it (I have seen other models around with 10TB SATA's).  It has two controllers, each controller has a 6gb cache module with a super capacitor.  The controllers each have four SPF ports, a SAS expander, a micro USB port for serial based CLI configuration, and 3.5mm headphone jack that is actually another serial connector that gets one into an ASCII driven menu.  

The Converged Network Adapters can operate in either 10Gb iSCSI mode, 16Gb Fibre Channel, or both!  I find that flexiblity a major plus; most vendors require a different controller all together.  The ports need to be configured to accept the type of GBIC installed.  Even after a firmware upgrade I was unable to make the changes via the web GUI. I did find it in the menuing via the ASCII menu. From the CLI one can change the charteristics,  in my case I set it to be in hybrid mode, where the first two ports are Fibre Channel and the second two are 10Gb iSCSI.  For both controllers, set all ports to use iSCSI protocol. # "set host-port-mode iSCSI"   For both controllers, set the first two ports to use FC protocol and the second two ports to use iSCSI protocol. # "set host-port-mode FC-and-iSCSI".

Firmware upgrades: the system must be error free before hand.  I had FC GBICs inserted in the controller when it was configured as iSCSI.  The system complained about the configuration profile, and the attempted firmware update left the system in a quasi failed state.  The primary controller took the upgrade but then appeared to be bricked.  Fortuntaly removing the offending GBIC and rebooting via the secondary controller allowed the updates to apply sucessfully.

Also another FYI, for some reason Putty would not work, despite trying several baud rate settings, with for serial based communication.  However TeraTerm works flawlessly.  No idea why.

ASCI menu on the left via the "head phone jack/Service-1" port; CLI via the USB port on the right.


Speed test connected to a Windows 2019 Server connected via 8Gb Qlogic FibreChannel

Speed test connected to a Windows 2019 Server connected via 16Gb Qlogic FibreChannel

Speed test connected to a Windows 2019 Server connected via 10gb iSCSI
Notice that the NIC utilization never even hits 6Gbps.  This is a single DAC cable going from SAN to server.

What Device manager shows when both FC ports are connected.

Rear of the unit showing both controllers. 

Each controller node contains three "special" DDR3 2Gb RAM modules, a supercap for keeping that cache memory alive during a power outage.  The OS seems to be loaded onto Compact Flash drives.  I am told the supercap keeps the data alive long enough to write the data to the CF card, and the OS is stored onto Non-volital RAM. 

Intel SR0NX Celeron 725c, single core CPU @ 1.3ghz

Default credientials are username: "manage" and password: "!manage"

TrueNAS revisited....iSCSI testing

As mentioned on previous posts, the new TrueNAS is a PowerEdge R320 w/ 48gb of ECC RAM, four HGST 10tb SATA 7200rpm hard drives, an NVMe drive, and a 10gb NIC.  The primary pupose of this box is to be a backup target, however it might get used as a shared iSCSI target from time to time, so it seemed worth while to do a bit of benchmarking.

The test client is a Windows 2012r2 Server, running as a VM that lives on VMware ESXi v7, installed on a HP ML350 G8 connected via a 10gb NIC.

For a baseline, here is the OS drive, that lives on three Seagate 6tb 7200rpm SAS drives, in a RAID5 configuration, connected via a HP Smart Array P420 with a 2gb cache.


TrueNAS was setup for iSCSI, a LUN created on the RAID10 SATA drives, connected to VMware, a datastore created, then an additional virtual disk created (thick provisioned, eager zero), and the same test ran.


TrueNAS was again setup with a LUN, but this time on the single PCI-e NVMe drive.  


Just for comparison, this is the very same Crucial NVMe drive as a local drive in a VMware server.

Here is a shot of what the 10gb NIC utilization looks like during each of the tests.  Interestingly the maximum transfer rate is roughly the same for each test, which would indicated we are reaching maximum network throughput; however the benchmark results would say there is still room for more speed.


FURTHER UPDATES:
Finally got around to swapping the E5-2450 v2 CPU for an E5-2418L v2 CPU; according to Intel's ARK page the maxim CPU wattages is 95w vs 50w.  At idle according to the Kill-A-Watt the usages went from 100watts down to 90watts, and according to the iDRAC from 94watts down to 70watts.  A quick iSCSI and SMB benchmarking shows no slow down in file transfers; not that I expected them to as I have file compression and dedupe turned off; so the CPU doesn't have much to do.


SuperMicro "Micro Cloud"

 

SuperMicro X11SSD-f System board.

Welcome to SuperMicro's "MicroCloud".  Think of it as a poor man's blade server, except this is a shared-nothing approach.    Well actually, the only thing shared between the nodes is power.  Unlike a Cisco or HP blade servers their is no shared networking, no boot profiles, no shared storage, or aligning up all of the firmware revisions between the blade's BIOS, the virtual NICs, the blades remote management, or any of the other components.  There are trades offs.  Since this is just like having a bunch of really small servers in a small space, the cabeling is, well awkward at best.  On the other hand there is no learning curve for learning all the ins-and-outs for blade chassis systems.

Each 3U chassis holds eight servers, each server has two 3.5" drives, two 1gbp NIC, IMPI (remote management), a PCIe slot, which has a 100gb NIC installed.

So here we have four MicroClouds, which means 32 compute nodes, each has 1gb connection to a switch dedicated for IMPI, a 1gb connection for normal networking, and two 100gb switches for connectivity.  It was decided, maybe not the best approach but it was what we decided on.  Two MicroClouds on the bottom, then 1gb switch for IMPI, then the 100gb switch, another 1gb switch  for LAN traffic that has a 10gb uplink, the thrid and fourth MicroClouds and another 100gb switch.


Here is a blade from an older MicroCloud; this one has X10sle-df motherboards that contain two Xeon 56XX series CPU's.


Notebook computers and charging

 Litium battieries don't last forever, and you might be shortening their life!  Turns out that if you use a notebook as your daily driver and have it plugged into the most of the time that battery is constantly being charged.  Charging a notebook creates heat, that heat and constant charging shortens the life of the battery.  Remember the exploding Samsung cell phones a few years ago?  

This one is from a 5th generation i5 Dell sub compact notebook that is now on it's 2nd battery.



Machines made with in the past few years started implenting an optional change that helps that battery life.  Basically the machine no longer charges to 100%.  It charges to a lesser amount and hovers there.  In the case of Microsoft Surface note when this option is turned on, the machine charges to a maximum of 50%.  The obvious downside is that if one then needs to run off of battery one doesn't get the benifit of a full chager.  I have noticed that the machine is notiable cooler.  

Instructions for Microsoft Surfaces can be found here.  Many normal machines: go into the BIOS, Power Management, and change the setting to "Primarily A/C Use".

TrueNAS re-visited.....

Time to build a new FreeNAS/TrueNAS server to be my backup target system.

The system to be used is a Lenovo TS140 workstation class machine.  It has an Intel Xeon E3-1226 v3 CPU.  It will probably exchanged to an Intel i-3 4160t CPU, as it has roughly half the wattage TDP.  The 8gb of DDR3 RAM was replaced w/ 16gb unbuffered ECC RAM.  An Intel i520 10gb NIC, a PCI-e to NVME and M.2 SATA SSD adapter was installed.  The boot drive is a very old Intel 64gb 2.5" SSD, a 512gb SATA M.2 drive, and a 512gb NVMe drives were added.  Interestingly the system does not seem to like 3rd generation NVMe drives; several were tried.  The NVMe drive that ended up being used is an older Samsung SM953 drive that seems to not have issues with older systems.  Some used HGST 10tb 7200rpm hard drives were acquired.  Initially they were set up in a mirror.  The Lenovo has a total of five SATA ports, so the M.2 SATA SSD is removed, four 10tb SATA drives is not an option.  Just for clarification, SATA0 goes to Intel 64gb boot drive, SATA1 goes to M.2 SATA SSD, SATA 2~5 feed the spinning hard disks.  

With two drives installed the system consumes roughly 45watts at idle.  Initial testing was not very scientific.   From a Windows VM, housed by a VMware server, that lives on NMVe storage, and on a 10gb network.  A folder was copied containing two 5gb ISO images, observed the min, max, and average speeds and timed with a stop watch  Yes I admit it, there is lots of room for errors and subjectivity.  The NAS was rebooted after each test to make sure the ZFS cache was empty.

Later a third drive was added, using RAIDz1 configuration, basically the same thing as RAID5.  Note one cannot easily migrate storage types, the existing array needed to be destroyed, then recreated with the three drives.  The wattage consumed in this configuration jumped up to 53watts. 


Mirrored SATA: Max: 330Mb/s, Min: 185Mb/s

NVMe: Max: 420Mb/s, Min: 350Mb/s, 45 seconds

M.2 SATA SSD: Max: 360 Mb/s, Min: 300Mb/s, 49 seconds

RAIDz1 SATA: Max 400Mb/s, Min: 290Mb/s, 65 seconds

Another system?  An opportunity to acquire a Dell PowerEdge R320 presented itself.  The advantages of this over the Lenovo is that it has four 3.5" hot swap trays.  Remote management via Dell's iDRAC system.  Dual power supplies, more ram (it comes with 48gb DDR3 ECC composed of three 16gb sticks, plus three more open slots, for a total of 6 vs. 4 on the Lenovo.  The CPU is an Intel E5-2640 v2.  Lower power versions of compatible CPU's might be worth acquiring later on.  

The PowerEdge has one SATA port, plus the SAS 8087 connector that feeds the four ports on the disk backplane.  The one single SATA port goes into a notebook optical drive to 2.5" hard drive adapter which houses another Intel 64gb SSD for booting.  However, this means I don't have any unused SATA ports to feed the M.2 SATA SSD drive.

With two of the same 10tb SATA drives, it consumes 70 watts, according to iDRAC, not measured with the Kill-A-Watt.  With four drives it jumped up to 84 watts. According to the Kill-A-Watt the four drive configution the NVMe drive consumes 100 watts at idle.  FWIW with the system off, but plugged in the machine consumes 16 watts of power.

This system came with a Perc H310 Mini Monolithic RAID card.  These cards do support hard drives in "non-RAID" mode. It is not entirely clear how that differs from normal drive, but a hard drive with such a configuration is directly readable in a normal PC and visa versa.  A bit of research shows that the SMART info from the drives will not be passed along to the OS.  

People have tried to flash these into "IT-Mode" (making it a normal SAS/SATA HBA), most of those same people end up bricking the card!  Art-Of-The-Server has figured out how to re-flash it properly; also sells them pre-flashed on eBay, for roughly $50.  

The motherboard has the same SAS connector as the H310.  The Perc was removed from the system and SAS/SATA cable moved the cable from the Perc card to the motherboard.  Windows 2019 was installed, benchmarked the I/O, and the results were pretty close.  So that being said for now there doesn't seem to be any reason to spend time attempting to re-flash the Perc into IT-Mode.  For the time being only SATA drives will be used.  One downside to using the onboard SATA controller is that iDRAC is completely ignorant of the drives in the system; one cannot log into iDRAC and verify drive health.

two Perc H310  Mini Monolith RAID cards

When configuring for four drives there is basically have two options.  RAIDz2, which is similar to RAID6, where it has two parity bits, meaning with four drives 50% of the capacity is available and two drives can go bad.  The other option is to do RAID 10, kind-of.....one has to create one Pool that contains two mirrors.   So the first two drives are mirrored, the 3rd and 4th drives are a mirror.  The two mirror groups are stripped.  Again 50% of the capacity is retained, possibly two drives can fail, but if the wrong two fail the whole set is trashed.

Mirrord SATA: AVG: 200Mb/s, Min: 185Mb/s

RAIDz1 SATA: AVG: 300Mb/s, Min: 220Mb/s

RAIDz2 SATA: AVG: 300Mb/s, Min: 220Mb/s

Dual Mirrors SATA: AVG: 350, Min: 250Mb/s

Windows VM during mirrored test

Windows VM during M.2 SSD test

Windows VM during RAIDz2 test

Windows VM during NVMe test

Network Card reporting during RAIDz1 test


Network Card reporting during NVMe test

The Lenovo and the Dell R320 can not truly be compared in terms of speed, as the systems were tested in different environments, different switching infrastructure, different test VMs, etc.  However it does give a rough idea.  Going forward: iSCSI targets will be setup, put a VM on that target and do an actual I/O test.  Also on the docket is to test with various forms of cache.  Since the primary purpose of this server is a for being a backup target, and MAYBE shared storage for VMware, the L2Arc cache does me no good.  It might be advantageous to have the log files live on faster storage; however using a 512gb NVMe drive is way over kill, since using a 10gb NIC it will never use more than 20gb, and the entire device must be allocated to the Log vDev.

Intel PCIe 3.0 NVMe drives

 

Ran across these so I had to test them out.

This one is 1.6tb of NVMe storage; it gets REALLY warm, and according to Intel's own docs it consumes 11.5watts at idle and 30w while writing, which explains why it is so hot!  Pretty impressive performance numbers considering it's age (circa 2015). Something else rather odd, is the card actually shows up as two 800gb drives.  In Linux it shows up as two different controllers, each controller having one namespace.

Intel Ark product specs

storagereview.com review





Pre-Built Windows VMs

Need to stand up one or more Windows VM for testing purposes?   Microsoft has got you covered.  They publish many Windows VMs that time-bomb in 90 days.  These are intended for those Web Dev's for testing sites.

Windows 11: This one is kinda fat in terms of storage usage as it has SQL and .Net items loaded.  It will auto login with no password.

https://developer.microsoft.com/en-us/windows/downloads/virtual-machines/


Windows 7 and Windows 10 v1809.  The Windows 10 one will NOT update past v1809.  The Windows 10 VM uses a preinstalled user named 'IEUser' with the password 'Passw0rd!'.

https://developer.microsoft.com/en-us/microsoft-edge/tools/vms/?WT.mc_id=-blog-scottha


***UPDATE***  It appears that Microsoft has pulled all but the Windows 11 VMs.




Six port network card

 When four ports isn't enough, use six?  I have seen dozens of quad port NICs but never a six port.

Silicom Pe2g6i35 use the Intel's i350 chipset.

https://www.silicom-usa.com/pr/server-adapters/networking-adapters/gigabit-ethernet-networking-server-adapters/pe2g6i35-server-adapter/




Enterprise SSD tear down

I had some failed drives that I was unable to warranty, so off to recycling they go!  I took a few pictures as I thought they were interesting; there is so much more going on in these than a standard desktop drive.
Notice the super caps!



 

RAID performance comparison

 Time to look at an age old question....when one only has a few drives, what RAID configuration is best for performance. Conventional wisdom is:

RAID0 is the fastest as two drives are load sharing.  Also known as "striping," the data across two drives, in theory two drives are splitting the load equally.

RAID1 is next fastest.  Also known as "mirroring:", all data is being written to both drives but either or both data can provide reads.  

Finally RAID 5 being the slowest.  In a three drive system the data is split in half, two drives get the data chunks and a third drive gets a calulated parity bit that can be used to reconstruct either of the data chunks.  Those three pieces are rotated among the drives in the member set.  The calulation of the data is what takes time, in theory more time than the previously mentioned methods.

First test: HP Proliant ML350 G6, SmartArray P410 with 1gb battery backed cache.  The hard drives are HGST 7200rpm SATA drives.  This machine is running VMware ESXi, the RAID volume was created and a virtual hard drive was created on the datastore for that RAID volume using thick provisioning.  The Atto test was run twice just to make sure the results were consistant.

 



Results:  In this test RAID5 out performs both RAID1 and RAID0 in both reads and writes.  RAID1 had better read performance than RAID0.  However RAID0 was but was faster than RAID1 in writes.


Second Test:  Dell PowerEdge R620, Perc Mini H310.  The hard drives are Intel Enterprise 200gb 3Gbps SSD's.
RAID0 (stripe)

RAID1 (mirror)

RAID5 (span with parity)...holly cow!  Do not do Raid5 w/ a Perc H310!

RAID5(span with parity)(RAID Controller w/ cache)
Results: RAID0 was much faster writes, but a tiny bit slower in reads when compared to RAID1.  Also in this test RAID5 was both faster than both RAID 0 and RAID1 in both reads & writes.  However it should be mentioned that the machine used for the 2nd R5 test was slightly different, it had a slower CPU (e5-2640 vs e5-2650) it also had a better RAID card (a Perc H710 vs H310).  The two RAID cards both have the LSI 2008 chipset, but I the H710 has 512mb cache and the H310 has 0mb.

Third test:  not apples to apples, but interesting.  The speeds of test number two were a bit low, so I tried a Samsung Evo 840 (CIRCA 2014).  The first test is on a Windows 10 PC, the 2nd test in the same Dell R620 as above, but in a RAID0 configuration.

Results: Early SSD's leave much to be desired in terms of performance.  The Evo 840's are TLC drives.

So it would seem that, atleast from these tests.....a three drive RAID5 out performs a two drive RAID0 or RAID1 configuration.

Brocade ICX 64XX network switch notes

The Brocade ICX 6430-24 is a 24 port plus four SPF (1gb) ports

The Brocade ICX 6450-224 is a 24 port plus four SPF+ (10gb) ports (albeit, only two 10gb ports are licensed from the factory)

The "p" at the end of the model means POE. The CLI is very close to Cisco's.

Brocade Datasheet

My re-setup notes:

Serial console cable into the switch, hit the reset button, during the boot process tap the "b" key a bunch of times to get to the boot loader.  Once there type: "no password", this disables the password for the next boot.  Then type "boot", and the switch will boot.  Once it boots:

enable

erase startup-config

reload

enable

config terminal

username admin privilege 0 password password

ip dhcp-client enable

crypto-ssl certificate generate

aaa authentication login default local

aaa authentication web-server default local

web-management https

wr mem




To set a static IP:

ip dhcp-client disable
vlan 1
router-interface ve 1
exit
interface ve 1
ip address 192.168.1.55/24
exit
write mem


As noted earlier, licencing.......the person in the above link has a way to re-write the serial number, and apply a license key.   FWIW this hardware is well over 10 years old, and Brocade has chosen to ignore this legacy hardware, as it is all End Of Life and End Of Support.