100mb/s speed limit, please help
-
Hi,
Sorry for the lack of information. I will lay things out a bit better here for you.- We are using fog 1.2.0
- The OS is Debian 7.9 Wheezy
- We are stuck at 100 Megabits for imaging. Imaging is capped at around 680 MegaBytes per Minute both ways, which equates to about 90Mbps. Adding more clients to image cuts this number between them.
- The FOG server is a physical desktop machine.
- The FOG server is connected with a CAT 5e cable to a gigabit switch.
I tried the IPerf test before, and confirmed that I am capable of getting gigabit (1000Mbps) speeds accross the network that I am using. I have also connected to the share the images are hosted and copied them to another computer. I was able to do this at 75-100MBps, which would be most definitely be in the gigabit speed range.
The next test I tried was FTP. I was able to FTP an image at around 50-80MBps. It is a little bit slower, but still is completely reasonable, and much faster than imaging speeds.
All of the clients I am trying to image are connected to a gigabit switch with a Cat5e cable. The server is on the same gigabit switch.
-
There is a new setting called bitrate for storage nodes in trunk version of FOG that may be missing a zero (I do not mean replication bandwidth)
-
-
@Quazz said:
There is a new setting called bitrate for storage nodes in trunk version of FOG that may be missing a zero (I do not mean replication bandwidth)
First the simple answer. The bitrate is only for moving files/images between the master node and a storage node. Not between the fog server and the target computers.
-
@stowtwe Thank you for the details. I see that you are using the 1.2.0 stable version and not a trunk (pre1.3.0) version. Is that correct?
We may need to get one of the devs to verify, but image deployment (partclone) should use nfs to move files to the target computers. Your testing with iperf and ftp show that the computer is capable of transferring at wire speeds (I would expect 50MB-80MB would be your cap because of internal computer restrictions. The 125MB/s is a theoretical max for GbE network).
So I wonder if there are inefficiencies for NFS on your version of debian. (sorry rhel person here) Isn’t wheezy a bit old?
-
@george1421 said:
@Quazz said:
There is a new setting called bitrate for storage nodes in trunk version of FOG that may be missing a zero (I do not mean replication bandwidth)
First the simple answer. The bitrate is only for moving files/images between the master node and a storage node. Not between the fog server and the target computers.
Wait, then what exactly is the difference between the replication bandwidth setting (which controls at what speeds images are allowed to replicate to other storage nodes) and the bitrate setting?
-
@george1421 Yes, I am on 1.2.0 stable. Would it be worthwhile to upgrade to trunk?
I will look into NFS. currently the images are on a drive that is in the FOG Server Tower, and I have that set up as the master node. There are no other nodes. These are the settings.
The switch I am using is an unmanaged Cisco switch. It has no configuration at all. I do not have access to any of the managed switches on the network, so I am trying to work around them the best I can by not running stuff directly through them. I have proxy DHCP configured on the FOG server, among other things.
-
It’s probably your compression settings then.
-
@stowtwe I’m not sure the trunk version will help you with this one.
Looking at your max clients, I’m surprised that it is two. On the default install of 1.2.0 from yesterday this value is 10.
@Wayne-Workman in version 1.2.0 there isn’t a compression setting from what I can see in 1.2.0 stable. So what ever the default value is, is what it is.
-
@george1421 Yes there is, in FOG Config -> FOG Settings -> FOG Boot Settings -> FOG_PIGZ_COMP
-
@Wayne-Workman mea culpa, I stand corrected
-
OK on my clean build of 1.2.0 the compression settings are 9
On my production instance that is on 1.2.0 trunk 5070 the compression is 6
On my dev instance that is on 1.2.0 trunk 5676 the compression is 6 (must be default) because I know I haven’t tweaked this system.
Changing this now shouldn’t have an impact since the image that is captured already will be compressed at the higher value. Something I don’t know is where does the decompression happen, I’m suspecting the client computer. If that’s the case the slower (cpu wise) the computer is the slower the transfer rate will be.
-
Compression and decompression happens on the host. The default compression value was changed from 9 to 6 because 6 is faster (in probably 99% of cases).
If the OP can change his compression to 6 and then re-upload the image, he might find that his image deploys much faster afterwards.
@ch3i posted a script that will change the compression of an image without deploying/capturing again. It does it all on the server. It’s in the forums somewhere here. It needs to go in the wiki. Somebody hash tag it if they find it.
-
I have an image with compression at 0 or 1 (I can’t recall what it was for sure), and an image with compression at 3, but both have about the same speeds. I have also tried compression 9 just to try the other side of things, and unsurprisingly, things were much slower.
Recapturing the image is not a huge deal for me, and I am open to try anything right now.
-
@stowtwe Try with it set to 6.
-
@stowtwe said:
I have an image with compression at 0 or 1 (I can’t recall what it was for sure), and an image with compression at 3, but both have about the same speeds. I have also tried compression 9 just to try the other side of things, and unsurprisingly, things were much slower.
Just to be clear here, an image with compression of 0 or 1 gave the same transfer speeds as a 3 or 9? I just want to make sure we are not chasing the wrong assumption.
-
Is the speed “limiting” on all systems? Have you attempting imaging in other locations?
From the sounds of things, things are working exactly as intended. This means, if there’s a 10/100 SWITCH between any of the GIG parts and your imaging to a system on the other side of that “middleman” switch, it would be giving you the same type of thing.
Speed isn’t always related to decompression/compression though it does play into it. Uploads would be the most often AFFECTED issue with compression as the CLient machine is not only transferring the data, but compressing it before it goes up the drain.
I think we need to trace where the 10/100 is starting at. It could even be as simple as an improper punch down.
-
All of our connections are gigabit running to the computers we are trying to image. The only connection that is 100Mb is the line running from the switch to our router, but that should be irrelevant in this case because the packets do not need to go through the router to reach the hosts we are imaging. This WOULD be a bottle neck if we were imaging to other rooms in the building.
Here is a picture (There is another switch before the router, but you get the idea):
@george1421 Compression 9 was much slower than the others. I have also just recently tried compression 3, and it yielded similar results to compression 1 or 0. I will be trying 6 next to see how it goes.
-
THat’s why we need to narrow what and when and where things are happening.
Just because you “know” doesn’t mean that it couldn’t be doing something unexpected.
Seeing as it consistently showing the same results, it leads me to believe there is a 10/100 connection somewhere between. Maybe the imaging computer is on a different subnet than the fog server? If it is it would have to pass through the router and back to reach the fog server to begin with.
-
Also,
Could you try a TOTALLY different system and image it to see if you see the same limiting factors?
What are the systems that you’re seeing get the 100mb/s issues (specifically)? Age, nic, bios, etc… as possible?
Are the host’s nic Gigabit or 10/100?
There’s a lot of variables, and testing the same systems over and over seems a bit much to try to find a good solution.
Maybe let’s take out the problem systems for now and find out if it is indeed network, or system.