iscsi targert Server
DELL 860 with 6/i
RAID0 WD GP 750G X2
Open - DSS Version: 5.0up60.7101.3511 32bit
iSCSI Disk lv0000
iscsiClient
IBM X3200 Q6600 with 8GB DRAM
Win2003
![]()
iscsi targert Server
DELL 860 with 6/i
RAID0 WD GP 750G X2
Open - DSS Version: 5.0up60.7101.3511 32bit
iSCSI Disk lv0000
iscsiClient
IBM X3200 Q6600 with 8GB DRAM
Win2003
![]()
Yes. iscsi lun is set in Write Back mode ,too.
In heavy loading situation ,I strongley recommend cheap fc card.
I am waiting for my emulex fc card and barcade fc switch.
I've been messing with that system, and here's the batch file I made to do the test (in the same directory as SQLIO.exe):
The "ping -n 2 127.0.0.1>nul" is in there only because there's not a "sleep" command included by default in windows.Code:ECHO off set /a mynumber=10 FOR /L %%G IN (1,1,5) DO CALL :in_fora ECHO Done... GOTO :eof :in_fora set /a myblock=1 ECHO d:\testfile.dat 1 0x0 %mynumber%>mypar.txt FOR /L %%H IN (1,1,5) DO CALL :in_forb set /a mynumber=%mynumber%*4 GOTO :eof :in_forb set /a myblock=%myblock%*4 set /a myout=1 FOR /L %%I IN (1,1,5) DO CALL :in_forc GOTO :eof :in_forc sqlio.exe -kW -s20 -frandom -o%myout% -b%myblock% -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresult.txt ping -n 2 127.0.0.1>nul set /a myout=%myout%*4 GOTO :eof
This batch file is equivalent, but is much longer:
And here's the results on Open-E's Todd's test machine using a 4Gb FC link on both ends and a 3ware RAID card (only a single drive, but with write cache enabled on both the disk and controller) with about 512MB of memory on the card and about 2GB of system ram on the DSS side (but, I'm not sure that 32-bit FC works with more than 1GB of cache per volume):Code:ECHO d:\testfile.dat 1 0x0 10>mypar.txt sqlio.exe -kW -s20 -frandom -o1 -b4 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul sqlio.exe -kW -s20 -frandom -o4 -b4 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul sqlio.exe -kW -s20 -frandom -o16 -b4 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul sqlio.exe -kW -s20 -frandom -o64 -b4 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul sqlio.exe -kW -s20 -frandom -o256 -b4 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul sqlio.exe -kW -s20 -frandom -o1 -b16 -BH -LS -Fmypar.txt|find /i "IOs/sec">>sqlresults.txt ping -n 2 127.0.0.1>nul ... etc., etc. ...
![]()
Open-E DSS (Target)
------------------------
Server: Industry standard
RAID Controller: Areca 1160 (PCI-X) 1GB cache
RAID Mode: 2x RAID10 (8 HDDs per RAID set) WB enabled
HDD: 16x Barracuda ES.2 SATA 500GB
FC HBA: 2x QLogic QLA2462 (4Gbps)
CPU: 2x AMD Opteron 275
RAM: 16GB
Open-E Version: 5.0.DB49000000.3278 (64Bit)
VMware ESX Host (Initiator)
--------------------------
Server: FSC RX300 S4
FC HBA: 2x Emulex LP1150 (4Gbps)
CPU: 2x Intel E5430
RAM: 32GB
VMware ESX Version: 3.5 Update 3 Enterprise
FC Environment
------------------
FC Switch: 2x EMC 5100
FC GBics: 32x 4G
We've currently 6 Open-E DSS servers and 6 VMware ESX Hosts. Multipathing is disabled.
The VMware ESX Hosts from which I've benchmarked has currently 10 servers running on different Open-E DSS servers. The Open-E DSS Storage where the benchmarked virtual machine was stored, has currently 10 servers running.
@thx0701: You've tested a file with a size of 8MB. This will be cached by the RAID controller and though your results are not comparable.
@Robotbeat: Same to you. You should test with files 2 to 4 times larger than your RAID controller's cache to get meaningful measurement results.