Netgate Discussion Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Search
    • Register
    • Login

    Playing with fq_codel in 2.4

    Scheduled Pinned Locked Moved Traffic Shaping
    1.1k Posts 123 Posters 1.6m Views
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Z
      zwck @tman222
      last edited by zwck

      @tman222
      Hey, Thanks mate. i guess i am aware of both methodologies, i am more wondering how do you find the proper settings to type in there. I read throught, and played around with, https://calomel.org/freebsd_network_tuning.html this guide. But could not see any difference.

      Also for people who want to play around with flent:
      quick installation guide for ubuntu 16+

      sudo apt update
      sudo apt upgrade
      sudo apt install git
      
      git clone https://github.com/HewlettPackard/netperf.git
      cd netperf
      sudo apt install texinfo
      sudo apt install iperf
      sudo apt-get install automake -y
      sudo apt install autoconf -y
      sudo apt install python-pip -y
      pip install netlib
      pip install cpp
      ./autogen.sh
      
      autoconf configure.ac > configure
      sudo chmod 755 configure
      ./configure --enable-demo
      make
      make install
      
      sudo add-apt-repository ppa:tohojo/flent
      sudo apt update
      sudo apt install flent
      
      
      flent rrul -p all_scaled -l 60 -H flent-london.bufferbloat.net -t no_shaper -o RRUL_no_shaper.png
      
      1 Reply Last reply Reply Quote 0
      • T
        tman222
        last edited by

        Hi @zwck

        It's a lot of trial and error (i.e. testing) to see what works best for your use case(s). Keep in mind that a lot of the guides you will find are for tuning host computers and some of those suggestions may not work well for a firewall appliance.

        One other site that I have gotten some helpful tuning info from has been the BSD Router Project, for example:
        https://bsdrp.net/documentation/technical_docs/performance

        Hope this helps.

        Z 1 Reply Last reply Reply Quote 0
        • S
          sciencetaco
          last edited by

          Is there any reason you folks can think of why when I run the flent rrul/rrul_noclassification, my download seems to top out at 40mb/s to netperf-west.bufferbloat.net. When I run "netperfrunner.sh" from the same host, i get the following:

          flent:
          alt text

          script:

          2018-10-10 08:59:19 Testing netperf-west.bufferbloat.net (ipv4) with 4 streams down and up while pinging gstatic.com. Takes about 60 seconds.
           Download:  150.21 Mbps
             Upload:  10.27 Mbps
            Latency: (in msec, 61 pings, 0.00% packet loss)
                Min: 29.343
              10pct: 33.824
             Median: 44.323
                Avg: 45.461
              90pct: 57.069
                Max: 74.273
          

          I'm applying the limiter via floating rules on WAN. I'm using codel+fq_codel set to 390mb/s down and 19mb/s up.

          I've seen some people incorporating their limiters via in/out pipe on the default lan allow rule - is there some consensus on which method is "best"? I've got a bunch of vlans off that interface - if i went this method, i'd need to include the in/out pipe on every default allow rule for each vlan?

          thank you for all you've managed to figure out and explain to me thus far.

          Z 1 Reply Last reply Reply Quote 0
          • Z
            zwck @sciencetaco
            last edited by zwck

            @sciencetaco

            I asked about this as well. some posts up dthat explains it, its actually 4x40Mbps ~ 160 and 4x3 ~ 12 Mbps (when you start flent with the option --gui you can check total download and upload values)
            Why it tops out at about half your speed limit is difficult to say, maybe hardware/line limitations from you or the host? I started setting up the codel params with extreme reduced speeds. i.e. 1gbit line limit, codel limiters set to 100Mbit.

            S 1 Reply Last reply Reply Quote 1
            • Z
              zwck @tman222
              last edited by

              @tman222 said in Playing with fq_codel in 2.4:

              Hi @zwck

              It's a lot of trial and error (i.e. testing) to see what works best for your use case(s). Keep in mind that a lot of the guides you will find are for tuning host computers and some of those suggestions may not work well for a firewall appliance.

              One other site that I have gotten some helpful tuning info from has been the BSD Router Project, for example:
              https://bsdrp.net/documentation/technical_docs/performance

              Hope this helps.

              I just quickly skimmed this section with the outcome:
              changing :

              machdep.hyperthreading_allowed="0" -> 24% increased performance
              net.inet.ip.fastforwarding=1 (useless since freebsd11)
              hw.igb.rxd or  hw.igb.txd -> decrease performance
              hw.igb.rx_process_limit=100 to -1 -> improvement, 1.7% 
              max_interrupt_rate from 8000 to 32000 -> no benefit
              Disabling LRO and TSO -> no impact
              
              T 1 Reply Last reply Reply Quote 0
              • S
                sciencetaco @zwck
                last edited by

                @zwck I don't get how that could be. Both tests use netperf, right? The two test outputs provided were from the same host. So odd.

                Z 1 Reply Last reply Reply Quote 0
                • Z
                  zwck @sciencetaco
                  last edited by zwck

                  @sciencetaco i probably don't understand you properly:

                  I think what you linked shows the same result - flent shows 150 down and about 10 up and your script output shows 150 down and 10 up

                  S 1 Reply Last reply Reply Quote 0
                  • S
                    sciencetaco @zwck
                    last edited by

                    @zwck I think I overlooked the 4x multiplier on flent in your original reply,my bad. This satisfied my brain's need for clarification. thank you!!

                    1 Reply Last reply Reply Quote 1
                    • D
                      dtaht
                      last edited by

                      At one level, I'm apologetic about the default rrul plot being so complicated. You get the most at a glance that way. You can certainly choose to output the totals plot instead, if that's what you want. My fear was that people would just look at that all the time instead of the more complicated one, and my other fear was that people wouldn't actually switch to using the gui to more fully analyze the data.

                      And my third fear was that people wouldn't use the other tests. You can test your
                      download or upload in isolation with either the tcp_download/tcp_upload test (simple) or do something more complicated like --te=download_streams=4 tcp_ndown . In the flent network I have not personally been able to stress the servers much past 100mbit, so here's YET ANOTHER TEST that goes to two servers:

                      flent -s .02 -x -H flent-fremont.bufferbloat.net -H flent-newark.bufferbloat.net -H flent-fremont.bufferbloat.net -H flent-fremont.bufferbloat.net -t 'whatever' rtt_fair4be

                      I put in more detailed fine grain sampling (-x -s .02).

                      But I have a feeling you are running out of cpu/interrupts/context switches.

                      1 Reply Last reply Reply Quote 0
                      • T
                        tibere86 @zwck
                        last edited by tibere86

                        @zwck said in Playing with fq_codel in 2.4:
                        "hw.igb.rxd or hw.igb.txd -> decrease performance"
                        What were the hw.igb.rxd and hw.igb.txd values you tried? pfSense default is 1024 for both I think.

                        Z 1 Reply Last reply Reply Quote 0
                        • D
                          dtaht
                          last edited by

                          btw: it would cheer me up if people would show their "before" plot, also. I should put one up of what my connection looks like without shaping on it.

                          I don't know how to make slow hardware faster... and if you encounter issues with shaping 300mbit, well, take a look at how much better things get if you just shape the upload, and call it a day. Can you still get 300mbit down without a shaper? Does latency improve if you just shape the out?

                          1 Reply Last reply Reply Quote 0
                          • D
                            dtaht
                            last edited by

                            So, this is what my last comcast modem looked like, without shaping.

                            0_1539192502963_rrul_-_sri-newcomcast-sanity-check-nosqm-2.png

                            0_1539192524603_rrul_be-2017-03-01T123958.188746.sri_newcomcast_sanity_check_nosqm_2.flent.gz

                            See how long the red download flow takes to get to parity? 20 seconds. Thats because it started just slightly late, and could not catch up with the other flows. This is what happens to any new flow (like, um, dns or tcp or...) when you have a flow already eating the link and your RTT climbs to 1 sec....

                            The upload flows are almost completely starved (1sec RTT!).

                            1 Reply Last reply Reply Quote 0
                            • D
                              dtaht
                              last edited by dtaht

                              using the squarewave test is fun too. Not doin that... But anyway, for comparison, I get about twice the upload performance and 15ms added latency on this hardware (an arm neon) running cake... and I'm running low on cpu here. (There's also other real traffic). Same cablemodem....

                              0_1539193568220_rrul_be_-_layer_cake_90mbit-2.png
                              After i get off this call I'll kill the download shaper and see what happens... but I'm in a call while I was doing this and nobody noticed... :)!

                              0_1539193422893_rrul_be-2018-10-10T103500.294739.layer_cake_90mbit.flent.gz

                              1 Reply Last reply Reply Quote 0
                              • D
                                dtaht
                                last edited by

                                ok, glutton for punishment. Shaped up only. I sure hope the rest of the world isn't as miserably overbuffered as comcast's CMTSes are.....

                                0_1539194108264_rrul_be_-_layer_cake_90mbit_uponly-1.png

                                0_1539194140099_rrul_be-2018-10-10T105101.635034.layer_cake_90mbit_uponly.flent.gz

                                I campaigned hard to get the cable industry to cut their CMTS buffering to a 100ms TOPs. So we're still suffering. pie on the modem is not enough. cheap arm and x86 hardware is not enough....

                                1 Reply Last reply Reply Quote 0
                                • Z
                                  zwck @tibere86
                                  last edited by

                                  @tibere86 look at the link tman posted, they did try 1024,2048,4096 as values.

                                  T 1 Reply Last reply Reply Quote 1
                                  • T
                                    tibere86 @zwck
                                    last edited by

                                    @zwck Thanks. I see it now.

                                    1 Reply Last reply Reply Quote 0
                                    • uptownVagrantU
                                      uptownVagrant
                                      last edited by

                                      @dtaht thanks for filing the icmp + limiter + NAT bug #9024. I added my note addressing the comment on the filter rules.

                                      I also did some testing with OPNsense 18.7.4 where the aforementioned bug is not present. I've noticed that both with and without NAT, latency is lower across OPNsense with regard to limiters + codel + fq_codel at 800mbit in my tests using the lab. (Codel, and fq_codel settings were same for both distros) Trying to track down what's different between the distros that may address this.

                                      0_1539221669545_RRUL_C2758_OPNsense18.7.4_800Mbit_ECN_t060_FQ_CoDel_CoDel_NAT.png
                                      0_1539221734101_rrul-2018-10-10T181639.321042.C2758_OPNsense18_7_4_800Mbit_ECN_NAT.flent.gz

                                      0_1539221691423_RRUL_C2758_pfSense2.4.4_800Mbit_ECN_t060_FQ_CoDel_CoDel_NoNAT.png
                                      0_1539221755472_rrul-2018-10-10T182751.251478.C2758_pfSense2_4_4_800Mbit_ECN_NoNAT.flent.gz

                                      And for posterity, here is a comparison of a Frontier FIOS connection without and with fq_codel shaping today.

                                      0_1539221825019_C2558_pfSense2.4.4_u10mbit_d50mbit_FIOS_noshape_t034.png
                                      0_1539222060906_rrul-2018-10-10T142122.571654.C2558_pfSense2_4_4_u10mbit_d50mbit_FIOS_noshape.flent.gz

                                      0_1539221830909_C2558_pfSense2.4.4_u10mbit_d48.5mbit_FIOS_t042.png
                                      0_1539222072770_rrul-2018-10-10T174938.401297.C2558_pfSense2_4_4_u10mbit_d48_5mbit_FIOS_shaped.flent.gz

                                      1 Reply Last reply Reply Quote 0
                                      • X
                                        xciter327
                                        last edited by xciter327

                                        I distinctly remember my graphs being flatter before the 2.4.4 update.

                                        edit: re-uploading the pic results in an "error"

                                        0_1539252489446_rrul-2018-10-11T115959.048554.300_100-with-masks.flent.gz

                                        I am testing against my own vps server. Limiters are applied on the LAN interface as interface rules.

                                        1 Reply Last reply Reply Quote 0
                                        • X
                                          xciter327
                                          last edited by xciter327

                                          Here is one with rules applied on the WAN via a floating rule.
                                          0_1539253348246_scaled-300-100-no-masks-WAN-rule.png

                                          0_1539253359216_rrul-2018-10-11T121532.991070.300_100-no-masks-WAN-rule.flent.gz

                                          FYI, the above results are on 1G symmetrical link. The test server is also on a 1G symmetrical link. The LAN is NAT-ted on a CARP address.

                                          D 1 Reply Last reply Reply Quote 0
                                          • S
                                            strangegopher
                                            last edited by

                                            My results are really bad. It could be that my modem is listed on that bad modem website

                                            0_1539253849276_rrul_-_2018-10-11_03^%26^%51.png

                                            I tweaked some settings but keep getting bad results.

                                            ipfw sched show
                                            00001: 181.000 Mbit/s    0 ms burst 0
                                            q65537  50 sl. 0 flows (1 buckets) sched 1 weight 0 lmax 0 pri 0 droptail
                                             sched 1 type FQ_CODEL flags 0x0 0 buckets 0 active
                                             FQ_CODEL target 8ms interval 80ms quantum 1518 limit 10240 flows 1024 NoECN
                                               Children flowsets: 1
                                            00002:  16.000 Mbit/s    0 ms burst 0
                                            q65538  50 sl. 0 flows (1 buckets) sched 2 weight 0 lmax 0 pri 0 droptail
                                             sched 2 type FQ_CODEL flags 0x0 0 buckets 0 active
                                             FQ_CODEL target 8ms interval 80ms quantum 1518 limit 10240 flows 1024 NoECN
                                               Children flowsets: 2
                                            
                                            D 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post
                                            Copyright 2025 Rubicon Communications LLC (Netgate). All rights reserved.