Netgate Discussion Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Search
    • Register
    • Login

    intel x520-da2 low throughput

    Scheduled Pinned Locked Moved Hardware
    20 Posts 3 Posters 3.5k Views
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • W
      websun
      last edited by websun

      Iperf vs iperf3 in my case provides the exact same results.
      Also, running -P does not seem to increase the process count, just the parallel connection, I can see is only one thread.

      808f594c-a8a8-4bb7-95a9-b4737cdf1320-image.png

      Leaving the gateways aside, this is the speed I am getting between host salt (vlan1) and host smtp (vlan2) with the troublesome gateways in between doing the routing and some filtering. Both are linux machines.

      [root@salt salt]# iperf3 -c smtp
      Connecting to host smtp, port 5201
      [  4] local 192.168.1.118 port 45280 connected to 192.168.2.25 port 5201
      [ ID] Interval           Transfer     Bandwidth       Retr  Cwnd
      [  4]   0.00-1.00   sec   142 MBytes  1.19 Gbits/sec    4    460 KBytes
      [  4]   1.00-2.00   sec   147 MBytes  1.23 Gbits/sec    0    660 KBytes
      [  4]   2.00-3.00   sec   181 MBytes  1.52 Gbits/sec    0    844 KBytes
      [  4]   3.00-4.00   sec   209 MBytes  1.75 Gbits/sec    0   1021 KBytes
      [  4]   4.00-5.00   sec   168 MBytes  1.41 Gbits/sec   19    868 KBytes
      [  4]   5.00-6.00   sec   180 MBytes  1.51 Gbits/sec    0   1015 KBytes
      [  4]   6.00-7.00   sec   186 MBytes  1.56 Gbits/sec    0   1.12 MBytes
      ^C[  4]   7.00-7.56   sec   126 MBytes  1.89 Gbits/sec    0   1.20 MBytes
      - - - - - - - - - - - - - - - - - - - - - - - - -
      [ ID] Interval           Transfer     Bandwidth       Retr
      [  4]   0.00-7.56   sec  1.31 GBytes  1.49 Gbits/sec   23             sender
      [  4]   0.00-7.56   sec  0.00 Bytes  0.00 bits/sec                  receiver
      iperf3: interrupt - the client has terminated
      
      [root@salt salt]# traceroute smtp
      traceroute to smtp (192.168.2.25), 30 hops max, 60 byte packets
       1  g1 (192.168.1.1)  0.091 ms  0.072 ms  0.059 ms
       2  192.168.2.25 (192.168.2.25)  0.397 ms  0.348 ms  0.340 ms
      [root@salt salt]#
      
      

      Iperf3 running on pfsense just shows me the same I get running between devices and so on. I don't think it's the cpu. Prob the driver or whatever setting the now one_foot_in_the_grave freebsd wants for a 10G card to work more than 20% of it's speed.

      1 Reply Last reply Reply Quote 0
      • stephenw10S
        stephenw10 Netgate Administrator
        last edited by

        The ix driver can load multiple queues to process multiple connections through it using multiple cores. It does that by default and you should see that in the top output. If you run iperf with more connections it can use that.

        The CPU is pretty ancient but I would still expect to see more that that through it.

        Steve

        1 Reply Last reply Reply Quote 0
        • W
          websun
          last edited by

          If I boot up a live linux on the same machine with ancient cpu, I get full speed with almost no cpu usage. That is my problem right now, unfortunately.
          If some people can get at least half of that 10G via pfsense, I am curious how.

          1 Reply Last reply Reply Quote 0
          • stephenw10S
            stephenw10 Netgate Administrator
            last edited by stephenw10

            Did you have any sort of firewall running in Linux when you tested that?

            If you disable pf in pfSense and are testing through it instead of to it I would expect to see 10G line rate from iperf.

            With pf enabled you won't but I'd expect to see 5-6Gbps without changing anything. Something is not right on your system there.

            I think last time I tested I was seeing 3-4Gbps using a Q6600 and that's quite a lot more ancient!

            Steve

            1 Reply Last reply Reply Quote 1
            • W
              websun
              last edited by

              I will disable PF and check that way but with low expectations. Will have to be done after hours on these.
              That was just a plain linux with iptables disabled. However, obviously that is a linux kernel with linux drivers. For example we have another set of gateways in the other office, with even crappier cpus (l5640) but running routeros (linux kernel and drivers) with much more firewall rules and yet we get around 7G.

              I am also not sure for example on these ix0 interfaces which is used exclusively for carp, how does PF limit the performance? it's just one rule there that allows everything on that specific interface since it's connected straight to the other one. Running even through that "unfiltered" interface does not get more than 2.3 Gb/s.

              1 Reply Last reply Reply Quote 0
              • stephenw10S
                stephenw10 Netgate Administrator
                last edited by

                Just having pf enabled on there introduces significant overhead. But I think you're right it's not just pf at work here, you should be seeing more throughput. I think you need to check the ix drive is creating multiple queues and using them. Try running vmstat -i during a test.

                Linux will almost always perform better to various extents on given hardware. But not that much better.

                Steve

                1 Reply Last reply Reply Quote 1
                • W
                  websun
                  last edited by

                  Seems like it does create about 8 queues as this is another set of boxes but just with one cpu. Still, I was not yet able to exceed 2.5Gb/s and this is on a straight connection no switch in between, no nat or filtering.
                  A little bit ridiculous...

                  1 Reply Last reply Reply Quote 0
                  • stephenw10S
                    stephenw10 Netgate Administrator
                    last edited by

                    That's iperf3 directly between the boxes with pfSense on? With multiple connections? 8 to use all the queues.

                    Steve

                    1 Reply Last reply Reply Quote 0
                    • W
                      websun
                      last edited by

                      Yes, unfurtunately. I've been playing with all sorts of ifconfig options for ix0 nic, to no avail. Some of them make matter worse but none of them better.
                      I believe I must accept the faith and blame freebsd. I can't spend days unfortunately troubleshooting this, the card/system should work out of the box.

                      1 Reply Last reply Reply Quote 0
                      • stephenw10S
                        stephenw10 Netgate Administrator
                        last edited by

                        I would certainly expect more from that CPU. You could try FreeBSD directly and see what performance you're seeing. Could be some obscure hardware/driver compatibility issue.

                        Steve

                        1 Reply Last reply Reply Quote 1
                        • W
                          websun
                          last edited by

                          Alright, more digging has been done.
                          Setting the mtu to 9000 gets me 9.90 Gbits/sec. Of course jumbo frames should make a different but sincerely not that much expected.

                          So, I went on with my investigation and created two vms on same server, installed pfsense with the configs from the troublesome ones. Using virtio for net, initially I was getting pretty much the same however enabling tso and lro on vtnet0 which was the "carp" interface pushed to 28 GB/s WITHOUT jumbo frames (because they are on the same machine, local software bridge)
                          However, this introduced new issues such as iperf3: error - unable to write to stream socket: Permission denied, which I am sure is some default limitations that need tweaking.

                          1 Reply Last reply Reply Quote 1
                          • First post
                            Last post
                          Copyright 2025 Rubicon Communications LLC (Netgate). All rights reserved.