Hello folks.
I started a research project a few days ago and I had multiple sets of input from these forums. But I also have a series of tools and calculators available at my disposal... all of which yield different results. So, even though this is an exercise in sizing and theory, it's hard to feel confident about using or creating a tool with so many different points of input.
So here goes… My project is to setup a small Gigabit Ethernet network (which doesn't exist in real life) and design a solid backup solution for about 20 servers. This part of the project is only concerned about the network transmission part of the backup scenario.
From all my research, I can say the one thing that is constant is that 1000BaseT (GbE) has a raw transmission threshold of 1 billion bits (1,000,000,000). Unfortunately, after that, everything sways. Between talk about MB, GB, MiB, GiB, etc... there is a lot to content with as far as opinions.
So this is the theory I came up with, step by step to follow the math and the logic...
1. 1000BaseT / 1GbE raw transmission capability is capped at 1,000,000,000 physical bits per second.
2. If you divide this number by 10, you get the number of bytes that can be transmitted. So why 10 and not 8? I am trying to account for the 8b-10b conversion algorithm where bits become encoded in the datagram frames. So if you send a bit (8 bytes) from one endpoint to another endpoint of a 1GbE network, what happens is the 8b-10b algorithm operates on the data and in effect, transmits 10 physical bits for every 8 bits of data. This yields 100,000,000 real bytes that can be truly transmitted.
3. Up-convert Bytes to KB with a multiple of 1,024, then KB to MB with a multiple of 1,024. You end up with 95.3674 MB/sec as the upper limit.
Simplified, my formula looks like this…
(1,000,000,000 / 10) / (1,024 * 1,024) = 95.3674 MB/sec
Some calculations stated here and around the internet indicate a different formula and result on the real amount of data that can be transmitted.
The first one is…
(1,000,000,000 / 8) / (1,024 * 1,024) = 119.2093 MB/sec
And the second one is…
(1,000,000,000 / 8) / (1,000 * 1,000) = 125 MB/sec
---
OK folks. Is my premise correct or are one of the other premises correct; network overhead aside? I need someone to let me know if I’m on the right track or not. Thanks in advance everyone.
I started a research project a few days ago and I had multiple sets of input from these forums. But I also have a series of tools and calculators available at my disposal... all of which yield different results. So, even though this is an exercise in sizing and theory, it's hard to feel confident about using or creating a tool with so many different points of input.
So here goes… My project is to setup a small Gigabit Ethernet network (which doesn't exist in real life) and design a solid backup solution for about 20 servers. This part of the project is only concerned about the network transmission part of the backup scenario.
From all my research, I can say the one thing that is constant is that 1000BaseT (GbE) has a raw transmission threshold of 1 billion bits (1,000,000,000). Unfortunately, after that, everything sways. Between talk about MB, GB, MiB, GiB, etc... there is a lot to content with as far as opinions.
So this is the theory I came up with, step by step to follow the math and the logic...
1. 1000BaseT / 1GbE raw transmission capability is capped at 1,000,000,000 physical bits per second.
2. If you divide this number by 10, you get the number of bytes that can be transmitted. So why 10 and not 8? I am trying to account for the 8b-10b conversion algorithm where bits become encoded in the datagram frames. So if you send a bit (8 bytes) from one endpoint to another endpoint of a 1GbE network, what happens is the 8b-10b algorithm operates on the data and in effect, transmits 10 physical bits for every 8 bits of data. This yields 100,000,000 real bytes that can be truly transmitted.
3. Up-convert Bytes to KB with a multiple of 1,024, then KB to MB with a multiple of 1,024. You end up with 95.3674 MB/sec as the upper limit.
Simplified, my formula looks like this…
(1,000,000,000 / 10) / (1,024 * 1,024) = 95.3674 MB/sec
Some calculations stated here and around the internet indicate a different formula and result on the real amount of data that can be transmitted.
The first one is…
(1,000,000,000 / 8) / (1,024 * 1,024) = 119.2093 MB/sec
And the second one is…
(1,000,000,000 / 8) / (1,000 * 1,000) = 125 MB/sec
---
OK folks. Is my premise correct or are one of the other premises correct; network overhead aside? I need someone to let me know if I’m on the right track or not. Thanks in advance everyone.