Netgate Discussion Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Search
    • Register
    • Login

    New PPPoE backend, some feedback

    Scheduled Pinned Locked Moved Development
    217 Posts 18 Posters 31.5k Views
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • M
      mr_nets @w0w
      last edited by mr_nets

      @w0w said in New PPPoE backend, some feedback:

      @dsl-ottawa
      Did you remove the net.isr.dispatch=deferred? See PPPoE with multi-queue NICs

      @w0w I did remove this value, shouldn't I ? I didn't see any change without.

      w0wW 1 Reply Last reply Reply Quote 1
      • w0wW
        w0w @mr_nets
        last edited by

        @mr_nets said in New PPPoE backend, some feedback:

        @w0w I did remove this value, shouldn't I ? I didn't see any change without.

        AFAIK it should be removed on the new backend.
        Do you have any other tunes enabled, flow control, tcp segmentation offload, LRO, no?
        Just guessing... I have never seen anything over 1Gig running PPPoE...

        M 1 Reply Last reply Reply Quote 0
        • M
          mr_nets @w0w
          last edited by

          @w0w said in New PPPoE backend, some feedback:

          @mr_nets said in New PPPoE backend, some feedback:

          @w0w I did remove this value, shouldn't I ? I didn't see any change without.

          AFAIK it should be removed on the new backend.
          Do you have any other tunes enabled, flow control, tcp segmentation offload, LRO, no?
          Just guessing... I have never seen anything over 1Gig running PPPoE...

          Fine, every offload setting are disabled as well. The only thing I didn't remove is Jumbo Frame on PPPoE (MTU 1500) since my ISP support that.

          1 Reply Last reply Reply Quote 1
          • stephenw10S
            stephenw10 Netgate Administrator
            last edited by

            What CPU usage are you seeing when you test? What about per core usage? I one core still pegged at 100%

            D M 2 Replies Last reply Reply Quote 0
            • D
              dsl.ottawa @stephenw10
              last edited by

              @stephenw10
              The multiple connection question is one I can't really answer. All I can say is that it's my providers tester and they sell up to 8gig on residential , so I have to sort of assume that it would take multiple streams into account. There is a comment on their page that mentions Ookla and Speedtest so there may be a linkage to them for the testing. It's an interesting thought though. Is there a known good tester that would be a good one to use to test against ? I'm betting it's going to be hard to get a full pipe test across the internet.

              Yup I've tried both cards for the WAN side of the connection.

              I've even thought that maybe the problem is on the PFSENSE side be elsewhere, like what if it's my desktop that is having the issue (long shot but the reality is that it is a device in the chain of the test)

              1 Reply Last reply Reply Quote 0
              • D
                dsl.ottawa @w0w
                last edited by

                @w0w
                I double checked and yup currently it is not there.
                so in my tunables side here are the ones I think I have changed (some have been there for a while to help the 10gig on the LAN side and general performance). and Some where attempts to fix this issue. I'm certainly not an expert when it comes to tuning so it's very possible that something is messed up.

                kern.ipc.somaxconn 4096
                hw.ix.tx_process_limit -1
                hw.ix.rx_process_limit -1
                hw.ix.rxd 4096
                hw.ix.txd 4096
                hw.ix.flow_control 0
                hw.intr_storm_threshold 10000
                net.isr.bindthreads 1
                net.isr.maxthreads 8
                net.inet.rss.enabled 1
                net.inet.rss.bits 2
                net.inet.ip.intr_queue_maxlen 4000
                net.isr.numthreads 8
                hw.ixl.tx_process_limit -1
                hw.ixl.rx_process_limit -1
                hw.ixl.rxd 4096
                hw.ixl.rxd 4096
                dev.ixl.0.fc 0
                dev.ixl.1.fc 0

                and in my boot local file

                net.inet.tcp.lro="0"
                hw.ixl.flow_control="0"
                hw.ix1.num_queues="8"
                dev.ixl.0.iflib.override_qs_enable=1
                dev.ixl.0.iflib.override_nrxqs=8
                dev.ixl.0.iflib.override_ntxqs=8
                dev.ixl.1.iflib.override_qs_enable=1
                dev.ixl.1.iflib.override_nrxqs=8
                dev.ixl.1.iflib.override_ntxqs=8
                dev.ixl.0.iflib.override_nrxds=4096
                dev.ixl.0.iflib.override_ntxds=4096
                dev.ixl.1.iflib.override_nrxds=4096
                dev.ixl.1.iflib.override_ntxds=4096

                And in the netowrk interface side

                TCP Segmentaion offloading, Large Receive and of course use if_pppoe are checked.

                more than 1gig pppoe, I'm not sure where in the world uses it but in Canada, they seem to love their PPPOE, we can get it up to 8gig from one of the providers. Honestly I still miss the old PVC based ATM connections, no special configuration from the users perspective it's just a pipe and it worked lol. I know the reason they use PPPOE it's just in cases like this it's extra overhead to deal with.

                1 Reply Last reply Reply Quote 1
                • stephenw10S
                  stephenw10 Netgate Administrator
                  last edited by

                  Yeah, if it's ookla it's almost certainly multistream. Also at those speeds it would probably need to be.

                  I'm not sure I can recommend anything other than the ISPs test at that speed either. Anything that was routed further may be restricted there.

                  But check the per-core CPU usage. Is it just hitting a single core limit?

                  1 Reply Last reply Reply Quote 0
                  • D
                    dsl.ottawa @stephenw10
                    last edited by

                    @stephenw10
                    the per core load doesn't seem to be an issue, which would mean that the new backend is doing it's job. I captured this in the middle of the download test.

                    CPU 0: 9.4% user, 0.0% nice, 43.1% system, 0.0% interrupt, 47.5% idle
                    CPU 1: 7.8% user, 0.0% nice, 1.6% system, 18.0% interrupt, 72.5% idle
                    CPU 2: 3.9% user, 0.0% nice, 3.1% system, 45.9% interrupt, 47.1% idle
                    CPU 3: 6.7% user, 0.0% nice, 3.5% system, 14.5% interrupt, 75.3% idle
                    CPU 4: 4.7% user, 0.0% nice, 14.5% system, 27.1% interrupt, 53.7% idle
                    CPU 5: 20.4% user, 0.0% nice, 7.1% system, 0.8% interrupt, 71.8% idle
                    CPU 6: 12.9% user, 0.0% nice, 28.6% system, 0.4% interrupt, 58.0% idle
                    CPU 7: 5.1% user, 0.0% nice, 2.7% system, 27.8% interrupt, 64.3% idle

                    1 Reply Last reply Reply Quote 0
                    • stephenw10S
                      stephenw10 Netgate Administrator
                      last edited by

                      Hmm, yeah that looks fine. Yet you still see all the traffic on one receive queue?

                      D 1 Reply Last reply Reply Quote 0
                      • w0wW
                        w0w
                        last edited by

                        Just FYI, some parameters used on one of my firewall with X710 and Xeon(R) CPU E3-1285L v4 :

                        net.inet.rss.bucket_mapping: 0:0 1:1 2:2 3:3 4:4 5:5 6:6 7:7 8:0 9:1 10:2 11:3 12:4 13:5 14:6 15:7
                        net.inet.rss.debug: 0
                        net.inet.rss.basecpu: 0
                        net.inet.rss.buckets: 16
                        net.inet.rss.maxcpus: 64
                        net.inet.rss.ncpus: 8
                        net.inet.rss.maxbits: 7
                        net.inet.rss.mask: 15
                        net.inet.rss.bits: 4
                        net.inet.rss.hashalgo: 2
                        hw.ixl.flow_control: 0
                        hw.ixl.tx_itr: 122
                        hw.ixl.rx_itr: 62
                        hw.ixl.shared_debug_mask: 0
                        hw.ixl.core_debug_mask: 0
                        hw.ixl.enable_head_writeback: 1
                        hw.ixl.enable_vf_loopback: 1
                        hw.ixl.i2c_access_method: 0
                        dev.ixl.0.wake: 0
                        dev.ixl.0.mac.xoff_recvd: 0
                        dev.ixl.0.mac.xoff_txd: 0
                        dev.ixl.0.mac.xon_recvd: 0
                        dev.ixl.0.mac.xon_txd: 0
                        dev.ixl.0.mac.tx_frames_big: 0
                        dev.ixl.0.mac.tx_frames_1024_1522: 21680695
                        dev.ixl.0.mac.tx_frames_512_1023: 14381471
                        dev.ixl.0.mac.tx_frames_256_511: 420044
                        dev.ixl.0.mac.tx_frames_128_255: 1330792
                        dev.ixl.0.mac.tx_frames_65_127: 2233359
                        dev.ixl.0.mac.tx_frames_64: 94083
                        dev.ixl.0.mac.checksum_errors: 0
                        dev.ixl.0.mac.rx_jabber: 0
                        dev.ixl.0.mac.rx_oversized: 0
                        dev.ixl.0.mac.rx_fragmented: 0
                        dev.ixl.0.mac.rx_undersize: 0
                        dev.ixl.0.mac.rx_frames_big: 0
                        dev.ixl.0.mac.rx_frames_1024_1522: 837784
                        dev.ixl.0.mac.rx_frames_512_1023: 105087
                        dev.ixl.0.mac.rx_frames_256_511: 526010
                        dev.ixl.0.mac.rx_frames_128_255: 801107
                        dev.ixl.0.mac.rx_frames_65_127: 17592164
                        dev.ixl.0.mac.rx_frames_64: 96374
                        dev.ixl.0.mac.rx_length_errors: 0
                        dev.ixl.0.mac.remote_faults: 0
                        dev.ixl.0.mac.local_faults: 1
                        dev.ixl.0.mac.illegal_bytes: 0
                        dev.ixl.0.mac.crc_errors: 0
                        dev.ixl.0.mac.bcast_pkts_txd: 25
                        dev.ixl.0.mac.mcast_pkts_txd: 1622
                        dev.ixl.0.mac.ucast_pkts_txd: 40138797
                        dev.ixl.0.mac.good_octets_txd: 39093159271
                        dev.ixl.0.mac.rx_discards: 0
                        dev.ixl.0.mac.bcast_pkts_rcvd: 2970
                        dev.ixl.0.mac.mcast_pkts_rcvd: 13177
                        dev.ixl.0.mac.ucast_pkts_rcvd: 19942379
                        dev.ixl.0.mac.good_octets_rcvd: 2847966747
                        dev.ixl.0.pf.txq03.itr: 122
                        dev.ixl.0.pf.txq03.bytes: 4629080532
                        dev.ixl.0.pf.txq03.packets: 5725464
                        dev.ixl.0.pf.txq03.mss_too_small: 0
                        dev.ixl.0.pf.txq03.tso: 0
                        dev.ixl.0.pf.txq02.itr: 122
                        dev.ixl.0.pf.txq02.bytes: 11637468273
                        dev.ixl.0.pf.txq02.packets: 10638943
                        dev.ixl.0.pf.txq02.mss_too_small: 0
                        dev.ixl.0.pf.txq02.tso: 0
                        dev.ixl.0.pf.txq01.itr: 122
                        dev.ixl.0.pf.txq01.bytes: 18307072164
                        dev.ixl.0.pf.txq01.packets: 18105770
                        dev.ixl.0.pf.txq01.mss_too_small: 0
                        dev.ixl.0.pf.txq01.tso: 0
                        dev.ixl.0.pf.txq00.itr: 122
                        dev.ixl.0.pf.txq00.bytes: 4364881017
                        dev.ixl.0.pf.txq00.packets: 5673826
                        dev.ixl.0.pf.txq00.mss_too_small: 0
                        dev.ixl.0.pf.txq00.tso: 0
                        dev.ixl.0.pf.rxq03.itr: 62
                        dev.ixl.0.pf.rxq03.desc_err: 0
                        dev.ixl.0.pf.rxq03.bytes: 0
                        dev.ixl.0.pf.rxq03.packets: 0
                        dev.ixl.0.pf.rxq03.irqs: 2153626
                        dev.ixl.0.pf.rxq02.itr: 62
                        dev.ixl.0.pf.rxq02.desc_err: 0
                        dev.ixl.0.pf.rxq02.bytes: 0
                        dev.ixl.0.pf.rxq02.packets: 0
                        dev.ixl.0.pf.rxq02.irqs: 2243131
                        dev.ixl.0.pf.rxq01.itr: 62
                        dev.ixl.0.pf.rxq01.desc_err: 0
                        dev.ixl.0.pf.rxq01.bytes: 0
                        dev.ixl.0.pf.rxq01.packets: 0
                        dev.ixl.0.pf.rxq01.irqs: 5679657
                        dev.ixl.0.pf.rxq00.itr: 62
                        dev.ixl.0.pf.rxq00.desc_err: 0
                        dev.ixl.0.pf.rxq00.bytes: 2766924779
                        dev.ixl.0.pf.rxq00.packets: 19948205
                        dev.ixl.0.pf.rxq00.irqs: 13195725
                        dev.ixl.0.pf.rx_errors: 0
                        dev.ixl.0.pf.bcast_pkts_txd: 25
                        dev.ixl.0.pf.mcast_pkts_txd: 281474976710649
                        dev.ixl.0.pf.ucast_pkts_txd: 40138797
                        dev.ixl.0.pf.good_octets_txd: 38931601458
                        dev.ixl.0.pf.rx_discards: 4294966174
                        dev.ixl.0.pf.bcast_pkts_rcvd: 2970
                        dev.ixl.0.pf.mcast_pkts_rcvd: 791
                        dev.ixl.0.pf.ucast_pkts_rcvd: 19941980
                        dev.ixl.0.pf.good_octets_rcvd: 2846521088
                        dev.ixl.0.admin_irq: 3
                        dev.ixl.0.link_active_on_if_down: 1
                        dev.ixl.0.eee.rx_lpi_count: 0
                        dev.ixl.0.eee.tx_lpi_count: 0
                        dev.ixl.0.eee.rx_lpi_status: 0
                        dev.ixl.0.eee.tx_lpi_status: 0
                        dev.ixl.0.eee.enable: 1
                        dev.ixl.0.fw_lldp: 1
                        dev.ixl.0.dynamic_tx_itr: 0
                        dev.ixl.0.dynamic_rx_itr: 0
                        dev.ixl.0.rx_itr: 62
                        dev.ixl.0.tx_itr: 122
                        dev.ixl.0.unallocated_queues: 380
                        dev.ixl.0.fw_version: fw 9.820.73026 api 1.15 nvm 9.20 etid 8000d966 oem 22.5632.9
                        dev.ixl.0.current_speed: 10 Gbps
                        dev.ixl.0.supported_speeds: 7
                        dev.ixl.0.advertise_speed: 7
                        dev.ixl.0.fc: 0
                        dev.ixl.0.iflib.rxq3.rxq_fl0.buf_size: 2048
                        dev.ixl.0.iflib.rxq3.rxq_fl0.credits: 128
                        dev.ixl.0.iflib.rxq3.rxq_fl0.cidx: 0
                        dev.ixl.0.iflib.rxq3.rxq_fl0.pidx: 128
                        dev.ixl.0.iflib.rxq3.cpu: 6
                        dev.ixl.0.iflib.rxq2.rxq_fl0.buf_size: 2048
                        dev.ixl.0.iflib.rxq2.rxq_fl0.credits: 128
                        dev.ixl.0.iflib.rxq2.rxq_fl0.cidx: 0
                        dev.ixl.0.iflib.rxq2.rxq_fl0.pidx: 128
                        dev.ixl.0.iflib.rxq2.cpu: 4
                        dev.ixl.0.iflib.rxq1.rxq_fl0.buf_size: 2048
                        dev.ixl.0.iflib.rxq1.rxq_fl0.credits: 128
                        dev.ixl.0.iflib.rxq1.rxq_fl0.cidx: 0
                        dev.ixl.0.iflib.rxq1.rxq_fl0.pidx: 128
                        dev.ixl.0.iflib.rxq1.cpu: 2
                        dev.ixl.0.iflib.rxq0.rxq_fl0.buf_size: 2048
                        dev.ixl.0.iflib.rxq0.rxq_fl0.credits: 1023
                        dev.ixl.0.iflib.rxq0.rxq_fl0.cidx: 685
                        dev.ixl.0.iflib.rxq0.rxq_fl0.pidx: 684
                        dev.ixl.0.iflib.rxq0.cpu: 0
                        dev.ixl.0.iflib.txq3.r_abdications: 0
                        dev.ixl.0.iflib.txq3.r_restarts: 0
                        dev.ixl.0.iflib.txq3.r_stalls: 0
                        dev.ixl.0.iflib.txq3.r_starts: 5719770
                        dev.ixl.0.iflib.txq3.r_drops: 0
                        dev.ixl.0.iflib.txq3.r_enqueues: 5733475
                        dev.ixl.0.iflib.txq3.ring_state: pidx_head: 1123 pidx_tail: 1123 cidx: 1123 state: IDLE
                        dev.ixl.0.iflib.txq3.txq_cleaned: 10975024
                        dev.ixl.0.iflib.txq3.txq_processed: 10975032
                        dev.ixl.0.iflib.txq3.txq_in_use: 9
                        dev.ixl.0.iflib.txq3.txq_cidx_processed: 824
                        dev.ixl.0.iflib.txq3.txq_cidx: 816
                        dev.ixl.0.iflib.txq3.txq_pidx: 825
                        dev.ixl.0.iflib.txq3.no_tx_dma_setup: 0
                        dev.ixl.0.iflib.txq3.txd_encap_efbig: 0
                        dev.ixl.0.iflib.txq3.tx_map_failed: 0
                        dev.ixl.0.iflib.txq3.no_desc_avail: 0
                        dev.ixl.0.iflib.txq3.mbuf_defrag_failed: 0
                        dev.ixl.0.iflib.txq3.m_pullups: 4578827
                        dev.ixl.0.iflib.txq3.mbuf_defrag: 0
                        dev.ixl.0.iflib.txq3.cpu: 6
                        dev.ixl.0.iflib.txq2.r_abdications: 0
                        dev.ixl.0.iflib.txq2.r_restarts: 0
                        dev.ixl.0.iflib.txq2.r_stalls: 0
                        dev.ixl.0.iflib.txq2.r_starts: 10633370
                        dev.ixl.0.iflib.txq2.r_drops: 0
                        dev.ixl.0.iflib.txq2.r_enqueues: 10653511
                        dev.ixl.0.iflib.txq2.ring_state: pidx_head: 1863 pidx_tail: 1863 cidx: 1863 state: IDLE
                        dev.ixl.0.iflib.txq2.txq_cleaned: 20820188
                        dev.ixl.0.iflib.txq2.txq_processed: 20820196
                        dev.ixl.0.iflib.txq2.txq_in_use: 8
                        dev.ixl.0.iflib.txq2.txq_cidx_processed: 228
                        dev.ixl.0.iflib.txq2.txq_cidx: 220
                        dev.ixl.0.iflib.txq2.txq_pidx: 228
                        dev.ixl.0.iflib.txq2.no_tx_dma_setup: 0
                        dev.ixl.0.iflib.txq2.txd_encap_efbig: 0
                        dev.ixl.0.iflib.txq2.tx_map_failed: 0
                        dev.ixl.0.iflib.txq2.no_desc_avail: 0
                        dev.ixl.0.iflib.txq2.mbuf_defrag_failed: 0
                        dev.ixl.0.iflib.txq2.m_pullups: 10415382
                        dev.ixl.0.iflib.txq2.mbuf_defrag: 0
                        dev.ixl.0.iflib.txq2.cpu: 4
                        dev.ixl.0.iflib.txq1.r_abdications: 0
                        dev.ixl.0.iflib.txq1.r_restarts: 0
                        dev.ixl.0.iflib.txq1.r_stalls: 0
                        dev.ixl.0.iflib.txq1.r_starts: 18108957
                        dev.ixl.0.iflib.txq1.r_drops: 0
                        dev.ixl.0.iflib.txq1.r_enqueues: 18135709
                        dev.ixl.0.iflib.txq1.ring_state: pidx_head: 0669 pidx_tail: 0669 cidx: 0669 state: IDLE
                        dev.ixl.0.iflib.txq1.txq_cleaned: 35744277
                        dev.ixl.0.iflib.txq1.txq_processed: 35744285
                        dev.ixl.0.iflib.txq1.txq_in_use: 9
                        dev.ixl.0.iflib.txq1.txq_cidx_processed: 541
                        dev.ixl.0.iflib.txq1.txq_cidx: 533
                        dev.ixl.0.iflib.txq1.txq_pidx: 542
                        dev.ixl.0.iflib.txq1.no_tx_dma_setup: 0
                        dev.ixl.0.iflib.txq1.txd_encap_efbig: 0
                        dev.ixl.0.iflib.txq1.tx_map_failed: 0
                        dev.ixl.0.iflib.txq1.no_desc_avail: 0
                        dev.ixl.0.iflib.txq1.mbuf_defrag_failed: 0
                        dev.ixl.0.iflib.txq1.m_pullups: 17863485
                        dev.ixl.0.iflib.txq1.mbuf_defrag: 0
                        dev.ixl.0.iflib.txq1.cpu: 2
                        dev.ixl.0.iflib.txq0.r_abdications: 0
                        dev.ixl.0.iflib.txq0.r_restarts: 0
                        dev.ixl.0.iflib.txq0.r_stalls: 0
                        dev.ixl.0.iflib.txq0.r_starts: 5665998
                        dev.ixl.0.iflib.txq0.r_drops: 0
                        dev.ixl.0.iflib.txq0.r_enqueues: 5676493
                        dev.ixl.0.iflib.txq0.ring_state: pidx_head: 1485 pidx_tail: 1485 cidx: 1485 state: IDLE
                        dev.ixl.0.iflib.txq0.txq_cleaned: 10681309
                        dev.ixl.0.iflib.txq0.txq_processed: 10681317
                        dev.ixl.0.iflib.txq0.txq_in_use: 8
                        dev.ixl.0.iflib.txq0.txq_cidx_processed: 997
                        dev.ixl.0.iflib.txq0.txq_cidx: 989
                        dev.ixl.0.iflib.txq0.txq_pidx: 997
                        dev.ixl.0.iflib.txq0.no_tx_dma_setup: 0
                        dev.ixl.0.iflib.txq0.txd_encap_efbig: 0
                        dev.ixl.0.iflib.txq0.tx_map_failed: 0
                        dev.ixl.0.iflib.txq0.no_desc_avail: 0
                        dev.ixl.0.iflib.txq0.mbuf_defrag_failed: 0
                        dev.ixl.0.iflib.txq0.m_pullups: 5253756
                        dev.ixl.0.iflib.txq0.mbuf_defrag: 0
                        dev.ixl.0.iflib.txq0.cpu: 0
                        dev.ixl.0.iflib.override_nrxds: 0
                        dev.ixl.0.iflib.override_ntxds: 0
                        dev.ixl.0.iflib.allocated_msix_vectors: 5
                        dev.ixl.0.iflib.use_extra_msix_vectors: 0
                        dev.ixl.0.iflib.use_logical_cores: 0
                        dev.ixl.0.iflib.separate_txrx: 0
                        dev.ixl.0.iflib.core_offset: 0
                        dev.ixl.0.iflib.tx_abdicate: 0
                        dev.ixl.0.iflib.rx_budget: 0
                        dev.ixl.0.iflib.disable_msix: 0
                        dev.ixl.0.iflib.override_qs_enable: 0
                        dev.ixl.0.iflib.override_nrxqs: 0
                        dev.ixl.0.iflib.override_ntxqs: 0
                        dev.ixl.0.iflib.driver_version: 2.3.3-k
                        dev.ixl.0.%iommu: rid=0x100
                        dev.ixl.0.%parent: pci1
                        dev.ixl.0.%pnpinfo: vendor=0x8086 device=0x15ff subvendor=0x8086 subdevice=0x0006 class=0x020000
                        dev.ixl.0.%location: slot=0 function=0 dbsf=pci0:1:0:0 handle=\_SB_.PCI0.PEG0.PEGP
                        dev.ixl.0.%driver: ixl
                        dev.ixl.0.%desc: Intel(R) Ethernet Controller X710 for 10GBASE-T - 2.3.3-k
                        dev.ixl.%parent:
                        ixl0@pci0:1:0:0:	class=0x020000 rev=0x02 hdr=0x00 vendor=0x8086 device=0x15ff subvendor=0x8086 subdevice=0x0006
                            vendor     = 'Intel Corporation'
                            device     = 'Ethernet Controller X710 for 10GBASE-T'
                            class      = network
                            subclass   = ethernet
                            bar   [10] = type Prefetchable Memory, range 64, base 0xfbe000000, size 16777216, enabled
                            bar   [1c] = type Prefetchable Memory, range 64, base 0xfbf018000, size 32768, enabled
                            cap 01[40] = powerspec 3  supports D0 D3  current D0
                            cap 05[50] = MSI supports 1 message, 64 bit, vector masks
                            cap 11[70] = MSI-X supports 129 messages, enabled
                                         Table in map 0x1c[0x0], PBA in map 0x1c[0x1000]
                            cap 10[a0] = PCI-Express 2 endpoint max data 128(2048) FLR
                                         max read 512
                                         link x8(x8) speed 8.0(8.0) ASPM L1(L1)
                            cap 03[e0] = VPD
                            ecap 0001[100] = AER 2 0 fatal 0 non-fatal 1 corrected
                            ecap 0003[140] = Serial 1 8477c9ffff9196b4
                            ecap 000e[150] = ARI 1
                            ecap 0010[160] = SR-IOV 1 IOV disabled, Memory Space disabled, ARI disabled
                                             0 VFs configured out of 32 supported
                                             First VF RID Offset 0x0110, VF RID Stride 0x0001
                                             VF Device ID 0x154c
                                             Page Sizes: 4096 (enabled), 8192, 65536, 262144, 1048576, 4194304
                            ecap 0017[1a0] = TPH Requester 1
                            ecap 000d[1b0] = ACS 1 Source Validation unavailable, Translation Blocking unavailable
                                             P2P Req Redirect unavailable, P2P Cmpl Redirect unavailable
                                             P2P Upstream Forwarding unavailable, P2P Egress Control unavailable
                                             P2P Direct Translated unavailable, Enhanced Capability unavailable
                            ecap 0019[1d0] = PCIe Sec 1 lane errors 0
                        

                        @dsl-ottawa
                        Is there a way to test pfSense WAN to LAN throughput without using PPPoE?
                        I don’t remember exactly what I set up back then, but it was something like an iPerf server on the WAN side (hooked to 10Gbit switch) with a 10 Gbit interface — and of course, the same on the LAN side, but client...

                        D 2 Replies Last reply Reply Quote 0
                        • D
                          dsl.ottawa @stephenw10
                          last edited by

                          @stephenw10 Yup yup

                          dev.ixl.0.pf.txq07.packets: 8401098
                          dev.ixl.0.pf.txq06.packets: 7316577
                          dev.ixl.0.pf.txq05.packets: 3704542
                          dev.ixl.0.pf.txq04.packets: 4271522
                          dev.ixl.0.pf.txq03.packets: 3784653
                          dev.ixl.0.pf.txq02.packets: 7219027
                          dev.ixl.0.pf.txq01.packets: 4471821
                          dev.ixl.0.pf.txq00.packets: 2694355
                          dev.ixl.0.pf.rxq07.packets: 0
                          dev.ixl.0.pf.rxq06.packets: 0
                          dev.ixl.0.pf.rxq05.packets: 0
                          dev.ixl.0.pf.rxq04.packets: 0
                          dev.ixl.0.pf.rxq03.packets: 0
                          dev.ixl.0.pf.rxq02.packets: 0
                          dev.ixl.0.pf.rxq01.packets: 0
                          dev.ixl.0.pf.rxq00.packets: 47313613

                          1 Reply Last reply Reply Quote 0
                          • D
                            dsl.ottawa @w0w
                            last edited by

                            @w0w Thanks for this I'll do some comparisons and see if there is something different.

                            What I might do this weekend, is push the pppoe back up to the provider modem and just run it as dhcp from there. I know it will be double NAT but if it's something between the PPPOE and the queuing that should then clear up. If I get better results and the queues are better then it narrows it down, and proves the rest of my network. I had already ran iperf between other devices and the firewall when I upgraded servers and switches to 10gig to try and tune everything, so it shouldn't be but I don't want to discount it without some proof either.

                            1 Reply Last reply Reply Quote 1
                            • D
                              dsl.ottawa @w0w
                              last edited by

                              @w0w Taking into account that I have 4 more queues than you did here's what I see are differences.

                              dev.ixl.0.eee.enable: 0
                              dev.ixl.0.unallocated_queues: 760 <- makes sense I have more queues enabled.
                              dev.ixl.0.fw_version: fw 6.81.49447 api 1.7 nvm 6.80 etid 80003d72 oem 18.4608.9
                              dev.ixl.0.supported_speeds: 6 <- probably due to the firmware difference.
                              dev.ixl.0.advertise_speed: 6 <- probably due to the firmware difference.
                              dev.ixl.0.iflib.override_nrxds: 4096 <-these are in my overrides so expected.
                              dev.ixl.0.iflib.override_ntxds: 4096 <-these are in my overrides so expected.
                              dev.ixl.0.iflib.allocated_msix_vectors: 9
                              dev.ixl.0.iflib.override_qs_enable: 1
                              dev.ixl.0.iflib.override_nrxqs: 8 <-related to my extra queue tests.
                              dev.ixl.0.iflib.override_ntxqs: 8 <-related to my extra queue tests.

                              so
                              a) I could try updating the firmware, never a bad thing
                              b) unsure about the
                              dev.ixl.0.eee.enable: 0
                              dev.ixl.0.iflib.allocated_msix_vectors: 9
                              dev.ixl.0.iflib.override_qs_enable: 1
                              and
                              c)Your queue outputs show that on the RX side only the one queue was active on yours too.

                              w0wW 1 Reply Last reply Reply Quote 0
                              • M
                                mr_nets @stephenw10
                                last edited by

                                @stephenw10 said in New PPPoE backend, some feedback:

                                What CPU usage are you seeing when you test? What about per core usage? I one core still pegged at 100%

                                Netgate 6100

                                Internet: 1G(1200 Mb/s) Down / 1G Up

                                3x Stream 4K 60FPS on 3 device
                                2x Download using Steam on 2 device
                                1x OpenVPN session open on the router but not very busy
                                2x Speed test (spot where it reach 1 gig) on 1 device

                                36736fcc-6978-4d1d-9a04-448869eef4f0-band.PNG

                                This is what it's look like in Grafana Cloud , it's not the most accurate I've seen but it's really close (I saw some spike at 80% for cpu 0 system on top "top -P" during the speed test ). What we can observe is that cpu 0 still very busy compare to the other cpu.

                                66b7e7b5-3c9f-4015-ba5d-fd91881ffb18-metric.PNG

                                1 Reply Last reply Reply Quote 1
                                • w0wW
                                  w0w @dsl.ottawa
                                  last edited by

                                  @dsl-ottawa said in New PPPoE backend, some feedback:

                                  a) I could try updating the firmware, never a bad thing
                                  b) unsure about the
                                  dev.ixl.0.eee.enable: 0
                                  dev.ixl.0.iflib.allocated_msix_vectors: 9
                                  dev.ixl.0.iflib.override_qs_enable: 1
                                  and
                                  c)Your queue outputs show that on the RX side only the one queue was active on yours too.

                                  I don't think those parameters are related to the problem — they seem like some OEM firmware/CPU settings, maybe.
                                  And about the queue... That’s a good find! I didn’t pay attention to it before. Hmm, interesting...
                                  I have a gigabit PPPoE connection, and the new backend gives me slightly better speeds compared to MPD5 with the deferred option. It almost saturates my ISP’s Ethernet link — over 900 Mbps — which I think is a great result for 20-year-old Cat5 cabling.

                                  D 1 Reply Last reply Reply Quote 0
                                  • D
                                    dsl.ottawa @w0w
                                    last edited by

                                    @w0w I would agree nothing there in the settings..

                                    BUT I found some more info.. and I should have kept to my gut feeling and done the further tests before pinpointing the PPPOE.

                                    I disabled my pppoe and ran straight IP, to my provider modem, so double nat'd.

                                    1. All the queues were used. as expected.
                                    2. my desktop speed tests DIDNT CHANGE, still slow downloads
                                    3. I found a copy of the cli speedtest package for pfsense and installed it.
                                      - download test from the pfsense box came back at full 3gig
                                      - upload was slower than normal but I'm less worried about that since my desktop tests were fine
                                    4. put PPPOE back, and ran the same tests.
                                      - download 3gig from pfsense using speedtest-cli
                                      - and normal 1.7-1.9 on my desktop

                                    This to me proves that it's not the PPPOE or even the wan side of pfsense but something on my bleepin LAN, or at a minimum the LAN side of pfsense.

                                    On the up side it does seem to prove that the single queue on the download isn't an issue :)

                                    w0wW 1 Reply Last reply Reply Quote 0
                                    • stephenw10S
                                      stephenw10 Netgate Administrator
                                      last edited by

                                      Ah that's interesting!

                                      What did the per-core CPU usage look like when downloading at 3G over PPPoE? Though the speedtest itself will be using a lot.

                                      D 1 Reply Last reply Reply Quote 0
                                      • D
                                        dsl.ottawa @stephenw10
                                        last edited by

                                        @stephenw10 it's a pretty quick test so I'm not sure it's a good capture

                                        CPU 0: 11.8% user, 0.0% nice, 28.6% system, 0.0% interrupt, 59.6% idle
                                        CPU 1: 3.9% user, 0.0% nice, 6.3% system, 10.2% interrupt, 79.6% idle
                                        CPU 2: 12.2% user, 0.0% nice, 7.8% system, 7.1% interrupt, 72.9% idle
                                        CPU 3: 5.9% user, 0.0% nice, 6.3% system, 11.0% interrupt, 76.9% idle
                                        CPU 4: 10.6% user, 0.0% nice, 7.8% system, 8.2% interrupt, 73.3% idle
                                        CPU 5: 4.7% user, 0.0% nice, 8.6% system, 12.9% interrupt, 73.7% idle
                                        CPU 6: 5.1% user, 0.0% nice, 6.7% system, 17.3% interrupt, 71.0% idle
                                        CPU 7: 4.3% user, 0.0% nice, 7.1% system, 11.0% interrupt, 77.6% idle

                                        but it certainly looks clean

                                        1 Reply Last reply Reply Quote 1
                                        • w0wW
                                          w0w @dsl.ottawa
                                          last edited by

                                          @dsl-ottawa said in New PPPoE backend, some feedback:

                                          put PPPOE back, and ran the same tests.

                                          • download 3gig from pfsense using speedtest-cli
                                          • and normal 1.7-1.9 on my desktop

                                          Is there any chance that some old limiters, rules, traffic shaping, or other configurations are still active?
                                          If this is a clean install, I would suspect a bug in pfSense

                                          1 Reply Last reply Reply Quote 0
                                          • M
                                            mr_nets @w0w
                                            last edited by mr_nets

                                            I make a script to monitor the queues of the parent interface IX2 and all data goes through the queue 0 on the RX side.

                                            BTW my ISP use a VLAN (IX2 --> VLAN40 --> PPPoE) maybe it doesn't help to manage the queues correctly...

                                            In 24 Hours
                                            5a0e6b2d-e486-477e-a21d-885af26de75a-image.png

                                            w0wW 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post
                                            Copyright 2025 Rubicon Communications LLC (Netgate). All rights reserved.