I can tell you that i have precisely this issue in Chicago. But the fact is
that for me it was a result of rate-limiting at the IP provider ATT. It is
not necessarily related directly, but senior citizens :-) may recall the
differential up/down bandwidth on ISDN.
At my last apartment i had fiber directly into my bedroom. Here it is over
copper to the building wiring. I took a 25% hit on bandwidth up and down. I
yelled at them for a rate reduction, but no dice.

On Sat, Jul 27, 2019, 8:24 AM Rainer Dorsch <m...@bokomoko.de> wrote:

> Hi,
>
> I have a stretch box configured as VLAN router (Cubox -i2ex). There is a
> drastic difference between the bandwidth of the uplink (VLAN1) and the
> downlinks (VLAN 2 to 7):
>
> On 192.168.7.1 (VLAN 7: eth0.7) I see arround 9 MB/s in a simple test:
> rd@home:~$ wget -O /dev/null http://fs/debian-9.3.0-amd64-netinst.iso
> [...] (9.08 MB/s)
> rd@home:~$
>
> On 192.168.0.30 (VLAN 1: eth0.1) is see less than 10%:
> rd@home:~$ wget -O /dev/null
> https://git.kernel.org/torvalds/t/linux-5.3-rc1.tar.gz
> --2019-07-27 14:46:38--
> https://git.kernel.org/torvalds/t/linux-5.3-rc1.tar.gz
> [...] (339KB/s)
>
> To prove that it has nothing to do with the uplink (there is a Fritzbox
> 6430)
> itself, I connected another  machine on same VLAN 1 (192.168.0.203). So
> overall, the network looks like this
>
>
> Internet
> |
> |
> Fritz-Box
> |
> |       192.168.0.203
> |--------------------------------------- x86
> |
> | 192.168.0.30
> |
> Cubox i
> |
> | 192.168.7.*
> |
>
> Note, the Cubox-i has only 1 physical interface, drawn are the virtual
> interface.
>
> The x86 machine reaches also a much higher network bandwidth:
>
> rd@h370:~/tmp.nobackup$ wget -O /dev/null
> https://git.kernel.org/torvalds/t/
> linux-5.3-rc1.tar.gz
> <https://git.kernel.org/torvalds/t/linux-5.3-rc1.tar.gz>
> [...] (5,49 MB/s)
> rd@h370:~/tmp.nobackup$
>
> I did run ifstat to confirm that there is no other traffic which consumes
> all the
> bandwidth:
>
> rd@home:/etc/shorewall$ ifstat -i eth0.1 1
>       eth0.1
>  KB/s in  KB/s out
>   495.00     16.34
>   417.14     16.03
>   484.33     16.53
>   384.80     11.96
>   393.33     12.67
>   632.59     17.68
>   607.90     17.91
>   354.39     12.00
>   678.58     20.97
>  1119.24     26.88
>  1185.20     27.27
>   925.91     21.84
>  1245.82     27.88
>   940.69     26.06
>  1023.72     26.89
>  1114.13     26.33
>   997.74     24.56
>   876.72     19.49
>  1167.56     27.73
>   906.41     24.30
>  1127.62     27.36
>   919.79     20.01
>   915.31     20.95
>   990.86     23.97
>  1119.22     26.94
>   905.54     26.40
>  1143.21     28.44
>  1096.15     26.98
>   924.62     24.89
>  1076.53     24.87
>  1004.04     23.99
>   811.11     23.13
>   983.71     24.46
>   885.05     23.19
>  1052.26     43.33
>  1230.55     37.11
>  1517.61     33.67
>   818.60     24.37
>  1057.24     26.63
>  1131.38     26.47
>  1278.43     30.12
>  1123.24     24.31
>   788.14     21.74
>   757.56     23.86
>  1135.29     27.91
>  1161.76     25.15
>  1465.32     32.04
>  1175.41     26.16
>  1371.36     31.56
>   811.73     21.70
>   540.97     16.91
>   381.78     20.95
>   306.44     13.07
>   378.02     12.93
>   603.65     16.67
>   418.31     16.35
>   393.71     16.46
>   479.89     15.05
>   436.74     13.85
>   395.96     12.05
>   476.41     14.51
>   470.09     18.23
>   322.02     12.20
>   427.35     14.33
>   464.25     14.39
>   404.19     11.09
>   999.41     24.39
>   931.23     24.89
>
> Also top does not show any obvious overload:
>
> rd@home:~$ top
> top - 14:51:53 up 34 days,  1:18,  3 users,  load average: 1.10, 1.11,
> 1.24
> Tasks: 177 total,   2 running, 125 sleeping,   0 stopped,   0 zombie
> %Cpu(s):  5.2 us,  5.2 sy,  0.0 ni, 87.9 id,  0.2 wa,  0.0 hi,  1.5 si,
> 0.0
> st
> KiB Mem :  2062764 total,    89308 free,   239480 used,  1733976
> buff/cache
> KiB Swap:  4726172 total,  4726172 free,        0 used.  1740396 avail
> Mem
>
>  PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND
>
>
> 13341 rd        20   0   12072   7196   3716 S   3.6  0.3   0:01.33 wget
>
>
>
> 7221 root      20   0   62236   3508   2468 S   2.9  0.2 394:36.44
> owserver
>
>
> 1257 rd        20   0    4036   2304   2052 S   2.0  0.1 945:39.61 reed-
> contact-mo
>
>
> 7687 rd        20   0    4712   2508   1596 S   2.0  0.1 571:54.03
> monitor.sh
>
>
> 13597 rd        20   0    7212   2900   2276 R   1.6  0.1   0:00.53 top
>
>
>
> 1040 asterisk  20   0  221968  58292  30316 S   1.3  2.8 565:37.76
> asterisk
>
>
>   17 root      20   0       0      0      0 S   0.7  0.0 434:16.89
> ksoftirqd/1
>
>
>   10 root      20   0       0      0      0 R   0.3  0.0  94:31.09
> rcu_sched
>
>
>  202 root      20   0   27144   6192   5760 S   0.3  0.3  89:57.07 systemd-
> journal
>
>
> 8678 root       0 -20       0      0      0 I   0.3  0.0   0:22.47 kworker/
> 2:2H-kb
>
>
> 21622 root      20   0       0      0      0 I   0.3  0.0   0:01.57
> kworker/
> u8:3-ev
>
>
> 32581 root       0 -20       0      0      0 I   0.3  0.0   0:18.03
> kworker/
> 0:1H-kb
>
>
>    1 root      20   0   26672   5236   3852 S   0.0  0.3   2:42.16
> systemd
>
>
>    2 root      20   0       0      0      0 S   0.0  0.0   0:07.14
> kthreadd
>
>
>    3 root       0 -20       0      0      0 I   0.0  0.0   0:00.00 rcu_gp
>
>
>
>    4 root       0 -20       0      0      0 I   0.0  0.0   0:00.00
> rcu_par_gp
>
>
>    8 root       0 -20       0      0      0 I   0.0  0.0   0:00.00
> mm_percpu_wq
>
>
>    9 root      20   0       0      0      0 S   0.0  0.0   4:43.82
> ksoftirqd/0
>
>
>   11 root      20   0       0      0      0 I   0.0  0.0   0:00.00 rcu_bh
>
>
>
>   12 root      rt   0       0      0      0 S   0.0  0.0   2:43.11
> migration/0
>
> So in summary:
> -> The uplink of the cubox to the internet is slow (<10% of available
> bandwidth)
> -> The cubox can run on the physical interface (the0) much higher traffic
> (as
> shown on VLAN 7)
> -> Another x86 host can run much higher traffic into the Internet
>
> Any idea on what could restrict the bandwidth on the Cubox uplink is very
> welcome. Also any ideas to diagnose the issue further would be useful for
> me.
>
> Many thanks
> Rainer
>
>
> --
> Rainer Dorsch
> http://bokomoko.de/
>
>
>

Reply via email to