Netgate Discussion Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Search
    • Register
    • Login

    New PPPoE backend, some feedback

    Development
    12
    153
    7.7k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • D
      dsl.ottawa @stephenw10
      last edited by

      @stephenw10
      the per core load doesn't seem to be an issue, which would mean that the new backend is doing it's job. I captured this in the middle of the download test.

      CPU 0: 9.4% user, 0.0% nice, 43.1% system, 0.0% interrupt, 47.5% idle
      CPU 1: 7.8% user, 0.0% nice, 1.6% system, 18.0% interrupt, 72.5% idle
      CPU 2: 3.9% user, 0.0% nice, 3.1% system, 45.9% interrupt, 47.1% idle
      CPU 3: 6.7% user, 0.0% nice, 3.5% system, 14.5% interrupt, 75.3% idle
      CPU 4: 4.7% user, 0.0% nice, 14.5% system, 27.1% interrupt, 53.7% idle
      CPU 5: 20.4% user, 0.0% nice, 7.1% system, 0.8% interrupt, 71.8% idle
      CPU 6: 12.9% user, 0.0% nice, 28.6% system, 0.4% interrupt, 58.0% idle
      CPU 7: 5.1% user, 0.0% nice, 2.7% system, 27.8% interrupt, 64.3% idle

      1 Reply Last reply Reply Quote 0
      • stephenw10S
        stephenw10 Netgate Administrator
        last edited by

        Hmm, yeah that looks fine. Yet you still see all the traffic on one receive queue?

        D 1 Reply Last reply Reply Quote 0
        • w0wW
          w0w
          last edited by

          Just FYI, some parameters used on one of my firewall with X710 and Xeon(R) CPU E3-1285L v4 :

          net.inet.rss.bucket_mapping: 0:0 1:1 2:2 3:3 4:4 5:5 6:6 7:7 8:0 9:1 10:2 11:3 12:4 13:5 14:6 15:7
          net.inet.rss.debug: 0
          net.inet.rss.basecpu: 0
          net.inet.rss.buckets: 16
          net.inet.rss.maxcpus: 64
          net.inet.rss.ncpus: 8
          net.inet.rss.maxbits: 7
          net.inet.rss.mask: 15
          net.inet.rss.bits: 4
          net.inet.rss.hashalgo: 2
          hw.ixl.flow_control: 0
          hw.ixl.tx_itr: 122
          hw.ixl.rx_itr: 62
          hw.ixl.shared_debug_mask: 0
          hw.ixl.core_debug_mask: 0
          hw.ixl.enable_head_writeback: 1
          hw.ixl.enable_vf_loopback: 1
          hw.ixl.i2c_access_method: 0
          dev.ixl.0.wake: 0
          dev.ixl.0.mac.xoff_recvd: 0
          dev.ixl.0.mac.xoff_txd: 0
          dev.ixl.0.mac.xon_recvd: 0
          dev.ixl.0.mac.xon_txd: 0
          dev.ixl.0.mac.tx_frames_big: 0
          dev.ixl.0.mac.tx_frames_1024_1522: 21680695
          dev.ixl.0.mac.tx_frames_512_1023: 14381471
          dev.ixl.0.mac.tx_frames_256_511: 420044
          dev.ixl.0.mac.tx_frames_128_255: 1330792
          dev.ixl.0.mac.tx_frames_65_127: 2233359
          dev.ixl.0.mac.tx_frames_64: 94083
          dev.ixl.0.mac.checksum_errors: 0
          dev.ixl.0.mac.rx_jabber: 0
          dev.ixl.0.mac.rx_oversized: 0
          dev.ixl.0.mac.rx_fragmented: 0
          dev.ixl.0.mac.rx_undersize: 0
          dev.ixl.0.mac.rx_frames_big: 0
          dev.ixl.0.mac.rx_frames_1024_1522: 837784
          dev.ixl.0.mac.rx_frames_512_1023: 105087
          dev.ixl.0.mac.rx_frames_256_511: 526010
          dev.ixl.0.mac.rx_frames_128_255: 801107
          dev.ixl.0.mac.rx_frames_65_127: 17592164
          dev.ixl.0.mac.rx_frames_64: 96374
          dev.ixl.0.mac.rx_length_errors: 0
          dev.ixl.0.mac.remote_faults: 0
          dev.ixl.0.mac.local_faults: 1
          dev.ixl.0.mac.illegal_bytes: 0
          dev.ixl.0.mac.crc_errors: 0
          dev.ixl.0.mac.bcast_pkts_txd: 25
          dev.ixl.0.mac.mcast_pkts_txd: 1622
          dev.ixl.0.mac.ucast_pkts_txd: 40138797
          dev.ixl.0.mac.good_octets_txd: 39093159271
          dev.ixl.0.mac.rx_discards: 0
          dev.ixl.0.mac.bcast_pkts_rcvd: 2970
          dev.ixl.0.mac.mcast_pkts_rcvd: 13177
          dev.ixl.0.mac.ucast_pkts_rcvd: 19942379
          dev.ixl.0.mac.good_octets_rcvd: 2847966747
          dev.ixl.0.pf.txq03.itr: 122
          dev.ixl.0.pf.txq03.bytes: 4629080532
          dev.ixl.0.pf.txq03.packets: 5725464
          dev.ixl.0.pf.txq03.mss_too_small: 0
          dev.ixl.0.pf.txq03.tso: 0
          dev.ixl.0.pf.txq02.itr: 122
          dev.ixl.0.pf.txq02.bytes: 11637468273
          dev.ixl.0.pf.txq02.packets: 10638943
          dev.ixl.0.pf.txq02.mss_too_small: 0
          dev.ixl.0.pf.txq02.tso: 0
          dev.ixl.0.pf.txq01.itr: 122
          dev.ixl.0.pf.txq01.bytes: 18307072164
          dev.ixl.0.pf.txq01.packets: 18105770
          dev.ixl.0.pf.txq01.mss_too_small: 0
          dev.ixl.0.pf.txq01.tso: 0
          dev.ixl.0.pf.txq00.itr: 122
          dev.ixl.0.pf.txq00.bytes: 4364881017
          dev.ixl.0.pf.txq00.packets: 5673826
          dev.ixl.0.pf.txq00.mss_too_small: 0
          dev.ixl.0.pf.txq00.tso: 0
          dev.ixl.0.pf.rxq03.itr: 62
          dev.ixl.0.pf.rxq03.desc_err: 0
          dev.ixl.0.pf.rxq03.bytes: 0
          dev.ixl.0.pf.rxq03.packets: 0
          dev.ixl.0.pf.rxq03.irqs: 2153626
          dev.ixl.0.pf.rxq02.itr: 62
          dev.ixl.0.pf.rxq02.desc_err: 0
          dev.ixl.0.pf.rxq02.bytes: 0
          dev.ixl.0.pf.rxq02.packets: 0
          dev.ixl.0.pf.rxq02.irqs: 2243131
          dev.ixl.0.pf.rxq01.itr: 62
          dev.ixl.0.pf.rxq01.desc_err: 0
          dev.ixl.0.pf.rxq01.bytes: 0
          dev.ixl.0.pf.rxq01.packets: 0
          dev.ixl.0.pf.rxq01.irqs: 5679657
          dev.ixl.0.pf.rxq00.itr: 62
          dev.ixl.0.pf.rxq00.desc_err: 0
          dev.ixl.0.pf.rxq00.bytes: 2766924779
          dev.ixl.0.pf.rxq00.packets: 19948205
          dev.ixl.0.pf.rxq00.irqs: 13195725
          dev.ixl.0.pf.rx_errors: 0
          dev.ixl.0.pf.bcast_pkts_txd: 25
          dev.ixl.0.pf.mcast_pkts_txd: 281474976710649
          dev.ixl.0.pf.ucast_pkts_txd: 40138797
          dev.ixl.0.pf.good_octets_txd: 38931601458
          dev.ixl.0.pf.rx_discards: 4294966174
          dev.ixl.0.pf.bcast_pkts_rcvd: 2970
          dev.ixl.0.pf.mcast_pkts_rcvd: 791
          dev.ixl.0.pf.ucast_pkts_rcvd: 19941980
          dev.ixl.0.pf.good_octets_rcvd: 2846521088
          dev.ixl.0.admin_irq: 3
          dev.ixl.0.link_active_on_if_down: 1
          dev.ixl.0.eee.rx_lpi_count: 0
          dev.ixl.0.eee.tx_lpi_count: 0
          dev.ixl.0.eee.rx_lpi_status: 0
          dev.ixl.0.eee.tx_lpi_status: 0
          dev.ixl.0.eee.enable: 1
          dev.ixl.0.fw_lldp: 1
          dev.ixl.0.dynamic_tx_itr: 0
          dev.ixl.0.dynamic_rx_itr: 0
          dev.ixl.0.rx_itr: 62
          dev.ixl.0.tx_itr: 122
          dev.ixl.0.unallocated_queues: 380
          dev.ixl.0.fw_version: fw 9.820.73026 api 1.15 nvm 9.20 etid 8000d966 oem 22.5632.9
          dev.ixl.0.current_speed: 10 Gbps
          dev.ixl.0.supported_speeds: 7
          dev.ixl.0.advertise_speed: 7
          dev.ixl.0.fc: 0
          dev.ixl.0.iflib.rxq3.rxq_fl0.buf_size: 2048
          dev.ixl.0.iflib.rxq3.rxq_fl0.credits: 128
          dev.ixl.0.iflib.rxq3.rxq_fl0.cidx: 0
          dev.ixl.0.iflib.rxq3.rxq_fl0.pidx: 128
          dev.ixl.0.iflib.rxq3.cpu: 6
          dev.ixl.0.iflib.rxq2.rxq_fl0.buf_size: 2048
          dev.ixl.0.iflib.rxq2.rxq_fl0.credits: 128
          dev.ixl.0.iflib.rxq2.rxq_fl0.cidx: 0
          dev.ixl.0.iflib.rxq2.rxq_fl0.pidx: 128
          dev.ixl.0.iflib.rxq2.cpu: 4
          dev.ixl.0.iflib.rxq1.rxq_fl0.buf_size: 2048
          dev.ixl.0.iflib.rxq1.rxq_fl0.credits: 128
          dev.ixl.0.iflib.rxq1.rxq_fl0.cidx: 0
          dev.ixl.0.iflib.rxq1.rxq_fl0.pidx: 128
          dev.ixl.0.iflib.rxq1.cpu: 2
          dev.ixl.0.iflib.rxq0.rxq_fl0.buf_size: 2048
          dev.ixl.0.iflib.rxq0.rxq_fl0.credits: 1023
          dev.ixl.0.iflib.rxq0.rxq_fl0.cidx: 685
          dev.ixl.0.iflib.rxq0.rxq_fl0.pidx: 684
          dev.ixl.0.iflib.rxq0.cpu: 0
          dev.ixl.0.iflib.txq3.r_abdications: 0
          dev.ixl.0.iflib.txq3.r_restarts: 0
          dev.ixl.0.iflib.txq3.r_stalls: 0
          dev.ixl.0.iflib.txq3.r_starts: 5719770
          dev.ixl.0.iflib.txq3.r_drops: 0
          dev.ixl.0.iflib.txq3.r_enqueues: 5733475
          dev.ixl.0.iflib.txq3.ring_state: pidx_head: 1123 pidx_tail: 1123 cidx: 1123 state: IDLE
          dev.ixl.0.iflib.txq3.txq_cleaned: 10975024
          dev.ixl.0.iflib.txq3.txq_processed: 10975032
          dev.ixl.0.iflib.txq3.txq_in_use: 9
          dev.ixl.0.iflib.txq3.txq_cidx_processed: 824
          dev.ixl.0.iflib.txq3.txq_cidx: 816
          dev.ixl.0.iflib.txq3.txq_pidx: 825
          dev.ixl.0.iflib.txq3.no_tx_dma_setup: 0
          dev.ixl.0.iflib.txq3.txd_encap_efbig: 0
          dev.ixl.0.iflib.txq3.tx_map_failed: 0
          dev.ixl.0.iflib.txq3.no_desc_avail: 0
          dev.ixl.0.iflib.txq3.mbuf_defrag_failed: 0
          dev.ixl.0.iflib.txq3.m_pullups: 4578827
          dev.ixl.0.iflib.txq3.mbuf_defrag: 0
          dev.ixl.0.iflib.txq3.cpu: 6
          dev.ixl.0.iflib.txq2.r_abdications: 0
          dev.ixl.0.iflib.txq2.r_restarts: 0
          dev.ixl.0.iflib.txq2.r_stalls: 0
          dev.ixl.0.iflib.txq2.r_starts: 10633370
          dev.ixl.0.iflib.txq2.r_drops: 0
          dev.ixl.0.iflib.txq2.r_enqueues: 10653511
          dev.ixl.0.iflib.txq2.ring_state: pidx_head: 1863 pidx_tail: 1863 cidx: 1863 state: IDLE
          dev.ixl.0.iflib.txq2.txq_cleaned: 20820188
          dev.ixl.0.iflib.txq2.txq_processed: 20820196
          dev.ixl.0.iflib.txq2.txq_in_use: 8
          dev.ixl.0.iflib.txq2.txq_cidx_processed: 228
          dev.ixl.0.iflib.txq2.txq_cidx: 220
          dev.ixl.0.iflib.txq2.txq_pidx: 228
          dev.ixl.0.iflib.txq2.no_tx_dma_setup: 0
          dev.ixl.0.iflib.txq2.txd_encap_efbig: 0
          dev.ixl.0.iflib.txq2.tx_map_failed: 0
          dev.ixl.0.iflib.txq2.no_desc_avail: 0
          dev.ixl.0.iflib.txq2.mbuf_defrag_failed: 0
          dev.ixl.0.iflib.txq2.m_pullups: 10415382
          dev.ixl.0.iflib.txq2.mbuf_defrag: 0
          dev.ixl.0.iflib.txq2.cpu: 4
          dev.ixl.0.iflib.txq1.r_abdications: 0
          dev.ixl.0.iflib.txq1.r_restarts: 0
          dev.ixl.0.iflib.txq1.r_stalls: 0
          dev.ixl.0.iflib.txq1.r_starts: 18108957
          dev.ixl.0.iflib.txq1.r_drops: 0
          dev.ixl.0.iflib.txq1.r_enqueues: 18135709
          dev.ixl.0.iflib.txq1.ring_state: pidx_head: 0669 pidx_tail: 0669 cidx: 0669 state: IDLE
          dev.ixl.0.iflib.txq1.txq_cleaned: 35744277
          dev.ixl.0.iflib.txq1.txq_processed: 35744285
          dev.ixl.0.iflib.txq1.txq_in_use: 9
          dev.ixl.0.iflib.txq1.txq_cidx_processed: 541
          dev.ixl.0.iflib.txq1.txq_cidx: 533
          dev.ixl.0.iflib.txq1.txq_pidx: 542
          dev.ixl.0.iflib.txq1.no_tx_dma_setup: 0
          dev.ixl.0.iflib.txq1.txd_encap_efbig: 0
          dev.ixl.0.iflib.txq1.tx_map_failed: 0
          dev.ixl.0.iflib.txq1.no_desc_avail: 0
          dev.ixl.0.iflib.txq1.mbuf_defrag_failed: 0
          dev.ixl.0.iflib.txq1.m_pullups: 17863485
          dev.ixl.0.iflib.txq1.mbuf_defrag: 0
          dev.ixl.0.iflib.txq1.cpu: 2
          dev.ixl.0.iflib.txq0.r_abdications: 0
          dev.ixl.0.iflib.txq0.r_restarts: 0
          dev.ixl.0.iflib.txq0.r_stalls: 0
          dev.ixl.0.iflib.txq0.r_starts: 5665998
          dev.ixl.0.iflib.txq0.r_drops: 0
          dev.ixl.0.iflib.txq0.r_enqueues: 5676493
          dev.ixl.0.iflib.txq0.ring_state: pidx_head: 1485 pidx_tail: 1485 cidx: 1485 state: IDLE
          dev.ixl.0.iflib.txq0.txq_cleaned: 10681309
          dev.ixl.0.iflib.txq0.txq_processed: 10681317
          dev.ixl.0.iflib.txq0.txq_in_use: 8
          dev.ixl.0.iflib.txq0.txq_cidx_processed: 997
          dev.ixl.0.iflib.txq0.txq_cidx: 989
          dev.ixl.0.iflib.txq0.txq_pidx: 997
          dev.ixl.0.iflib.txq0.no_tx_dma_setup: 0
          dev.ixl.0.iflib.txq0.txd_encap_efbig: 0
          dev.ixl.0.iflib.txq0.tx_map_failed: 0
          dev.ixl.0.iflib.txq0.no_desc_avail: 0
          dev.ixl.0.iflib.txq0.mbuf_defrag_failed: 0
          dev.ixl.0.iflib.txq0.m_pullups: 5253756
          dev.ixl.0.iflib.txq0.mbuf_defrag: 0
          dev.ixl.0.iflib.txq0.cpu: 0
          dev.ixl.0.iflib.override_nrxds: 0
          dev.ixl.0.iflib.override_ntxds: 0
          dev.ixl.0.iflib.allocated_msix_vectors: 5
          dev.ixl.0.iflib.use_extra_msix_vectors: 0
          dev.ixl.0.iflib.use_logical_cores: 0
          dev.ixl.0.iflib.separate_txrx: 0
          dev.ixl.0.iflib.core_offset: 0
          dev.ixl.0.iflib.tx_abdicate: 0
          dev.ixl.0.iflib.rx_budget: 0
          dev.ixl.0.iflib.disable_msix: 0
          dev.ixl.0.iflib.override_qs_enable: 0
          dev.ixl.0.iflib.override_nrxqs: 0
          dev.ixl.0.iflib.override_ntxqs: 0
          dev.ixl.0.iflib.driver_version: 2.3.3-k
          dev.ixl.0.%iommu: rid=0x100
          dev.ixl.0.%parent: pci1
          dev.ixl.0.%pnpinfo: vendor=0x8086 device=0x15ff subvendor=0x8086 subdevice=0x0006 class=0x020000
          dev.ixl.0.%location: slot=0 function=0 dbsf=pci0:1:0:0 handle=\_SB_.PCI0.PEG0.PEGP
          dev.ixl.0.%driver: ixl
          dev.ixl.0.%desc: Intel(R) Ethernet Controller X710 for 10GBASE-T - 2.3.3-k
          dev.ixl.%parent:
          ixl0@pci0:1:0:0:	class=0x020000 rev=0x02 hdr=0x00 vendor=0x8086 device=0x15ff subvendor=0x8086 subdevice=0x0006
              vendor     = 'Intel Corporation'
              device     = 'Ethernet Controller X710 for 10GBASE-T'
              class      = network
              subclass   = ethernet
              bar   [10] = type Prefetchable Memory, range 64, base 0xfbe000000, size 16777216, enabled
              bar   [1c] = type Prefetchable Memory, range 64, base 0xfbf018000, size 32768, enabled
              cap 01[40] = powerspec 3  supports D0 D3  current D0
              cap 05[50] = MSI supports 1 message, 64 bit, vector masks
              cap 11[70] = MSI-X supports 129 messages, enabled
                           Table in map 0x1c[0x0], PBA in map 0x1c[0x1000]
              cap 10[a0] = PCI-Express 2 endpoint max data 128(2048) FLR
                           max read 512
                           link x8(x8) speed 8.0(8.0) ASPM L1(L1)
              cap 03[e0] = VPD
              ecap 0001[100] = AER 2 0 fatal 0 non-fatal 1 corrected
              ecap 0003[140] = Serial 1 8477c9ffff9196b4
              ecap 000e[150] = ARI 1
              ecap 0010[160] = SR-IOV 1 IOV disabled, Memory Space disabled, ARI disabled
                               0 VFs configured out of 32 supported
                               First VF RID Offset 0x0110, VF RID Stride 0x0001
                               VF Device ID 0x154c
                               Page Sizes: 4096 (enabled), 8192, 65536, 262144, 1048576, 4194304
              ecap 0017[1a0] = TPH Requester 1
              ecap 000d[1b0] = ACS 1 Source Validation unavailable, Translation Blocking unavailable
                               P2P Req Redirect unavailable, P2P Cmpl Redirect unavailable
                               P2P Upstream Forwarding unavailable, P2P Egress Control unavailable
                               P2P Direct Translated unavailable, Enhanced Capability unavailable
              ecap 0019[1d0] = PCIe Sec 1 lane errors 0
          

          @dsl-ottawa
          Is there a way to test pfSense WAN to LAN throughput without using PPPoE?
          I don’t remember exactly what I set up back then, but it was something like an iPerf server on the WAN side (hooked to 10Gbit switch) with a 10 Gbit interface — and of course, the same on the LAN side, but client...

          D 2 Replies Last reply Reply Quote 0
          • D
            dsl.ottawa @stephenw10
            last edited by

            @stephenw10 Yup yup

            dev.ixl.0.pf.txq07.packets: 8401098
            dev.ixl.0.pf.txq06.packets: 7316577
            dev.ixl.0.pf.txq05.packets: 3704542
            dev.ixl.0.pf.txq04.packets: 4271522
            dev.ixl.0.pf.txq03.packets: 3784653
            dev.ixl.0.pf.txq02.packets: 7219027
            dev.ixl.0.pf.txq01.packets: 4471821
            dev.ixl.0.pf.txq00.packets: 2694355
            dev.ixl.0.pf.rxq07.packets: 0
            dev.ixl.0.pf.rxq06.packets: 0
            dev.ixl.0.pf.rxq05.packets: 0
            dev.ixl.0.pf.rxq04.packets: 0
            dev.ixl.0.pf.rxq03.packets: 0
            dev.ixl.0.pf.rxq02.packets: 0
            dev.ixl.0.pf.rxq01.packets: 0
            dev.ixl.0.pf.rxq00.packets: 47313613

            1 Reply Last reply Reply Quote 0
            • D
              dsl.ottawa @w0w
              last edited by

              @w0w Thanks for this I'll do some comparisons and see if there is something different.

              What I might do this weekend, is push the pppoe back up to the provider modem and just run it as dhcp from there. I know it will be double NAT but if it's something between the PPPOE and the queuing that should then clear up. If I get better results and the queues are better then it narrows it down, and proves the rest of my network. I had already ran iperf between other devices and the firewall when I upgraded servers and switches to 10gig to try and tune everything, so it shouldn't be but I don't want to discount it without some proof either.

              1 Reply Last reply Reply Quote 1
              • D
                dsl.ottawa @w0w
                last edited by

                @w0w Taking into account that I have 4 more queues than you did here's what I see are differences.

                dev.ixl.0.eee.enable: 0
                dev.ixl.0.unallocated_queues: 760 <- makes sense I have more queues enabled.
                dev.ixl.0.fw_version: fw 6.81.49447 api 1.7 nvm 6.80 etid 80003d72 oem 18.4608.9
                dev.ixl.0.supported_speeds: 6 <- probably due to the firmware difference.
                dev.ixl.0.advertise_speed: 6 <- probably due to the firmware difference.
                dev.ixl.0.iflib.override_nrxds: 4096 <-these are in my overrides so expected.
                dev.ixl.0.iflib.override_ntxds: 4096 <-these are in my overrides so expected.
                dev.ixl.0.iflib.allocated_msix_vectors: 9
                dev.ixl.0.iflib.override_qs_enable: 1
                dev.ixl.0.iflib.override_nrxqs: 8 <-related to my extra queue tests.
                dev.ixl.0.iflib.override_ntxqs: 8 <-related to my extra queue tests.

                so
                a) I could try updating the firmware, never a bad thing
                b) unsure about the
                dev.ixl.0.eee.enable: 0
                dev.ixl.0.iflib.allocated_msix_vectors: 9
                dev.ixl.0.iflib.override_qs_enable: 1
                and
                c)Your queue outputs show that on the RX side only the one queue was active on yours too.

                w0wW 1 Reply Last reply Reply Quote 0
                • M
                  mr_nets @stephenw10
                  last edited by

                  @stephenw10 said in New PPPoE backend, some feedback:

                  What CPU usage are you seeing when you test? What about per core usage? I one core still pegged at 100%

                  Netgate 6100

                  Internet: 1G(1200 Mb/s) Down / 1G Up

                  3x Stream 4K 60FPS on 3 device
                  2x Download using Steam on 2 device
                  1x OpenVPN session open on the router but not very busy
                  2x Speed test (spot where it reach 1 gig) on 1 device

                  36736fcc-6978-4d1d-9a04-448869eef4f0-band.PNG

                  This is what it's look like in Grafana Cloud , it's not the most accurate I've seen but it's really close (I saw some spike at 80% for cpu 0 system on top "top -P" during the speed test ). What we can observe is that cpu 0 still very busy compare to the other cpu.

                  66b7e7b5-3c9f-4015-ba5d-fd91881ffb18-metric.PNG

                  1 Reply Last reply Reply Quote 1
                  • w0wW
                    w0w @dsl.ottawa
                    last edited by

                    @dsl-ottawa said in New PPPoE backend, some feedback:

                    a) I could try updating the firmware, never a bad thing
                    b) unsure about the
                    dev.ixl.0.eee.enable: 0
                    dev.ixl.0.iflib.allocated_msix_vectors: 9
                    dev.ixl.0.iflib.override_qs_enable: 1
                    and
                    c)Your queue outputs show that on the RX side only the one queue was active on yours too.

                    I don't think those parameters are related to the problem — they seem like some OEM firmware/CPU settings, maybe.
                    And about the queue... That’s a good find! I didn’t pay attention to it before. Hmm, interesting...
                    I have a gigabit PPPoE connection, and the new backend gives me slightly better speeds compared to MPD5 with the deferred option. It almost saturates my ISP’s Ethernet link — over 900 Mbps — which I think is a great result for 20-year-old Cat5 cabling.

                    D 1 Reply Last reply Reply Quote 0
                    • D
                      dsl.ottawa @w0w
                      last edited by

                      @w0w I would agree nothing there in the settings..

                      BUT I found some more info.. and I should have kept to my gut feeling and done the further tests before pinpointing the PPPOE.

                      I disabled my pppoe and ran straight IP, to my provider modem, so double nat'd.

                      1. All the queues were used. as expected.
                      2. my desktop speed tests DIDNT CHANGE, still slow downloads
                      3. I found a copy of the cli speedtest package for pfsense and installed it.
                        - download test from the pfsense box came back at full 3gig
                        - upload was slower than normal but I'm less worried about that since my desktop tests were fine
                      4. put PPPOE back, and ran the same tests.
                        - download 3gig from pfsense using speedtest-cli
                        - and normal 1.7-1.9 on my desktop

                      This to me proves that it's not the PPPOE or even the wan side of pfsense but something on my bleepin LAN, or at a minimum the LAN side of pfsense.

                      On the up side it does seem to prove that the single queue on the download isn't an issue :)

                      w0wW 1 Reply Last reply Reply Quote 0
                      • stephenw10S
                        stephenw10 Netgate Administrator
                        last edited by

                        Ah that's interesting!

                        What did the per-core CPU usage look like when downloading at 3G over PPPoE? Though the speedtest itself will be using a lot.

                        D 1 Reply Last reply Reply Quote 0
                        • D
                          dsl.ottawa @stephenw10
                          last edited by

                          @stephenw10 it's a pretty quick test so I'm not sure it's a good capture

                          CPU 0: 11.8% user, 0.0% nice, 28.6% system, 0.0% interrupt, 59.6% idle
                          CPU 1: 3.9% user, 0.0% nice, 6.3% system, 10.2% interrupt, 79.6% idle
                          CPU 2: 12.2% user, 0.0% nice, 7.8% system, 7.1% interrupt, 72.9% idle
                          CPU 3: 5.9% user, 0.0% nice, 6.3% system, 11.0% interrupt, 76.9% idle
                          CPU 4: 10.6% user, 0.0% nice, 7.8% system, 8.2% interrupt, 73.3% idle
                          CPU 5: 4.7% user, 0.0% nice, 8.6% system, 12.9% interrupt, 73.7% idle
                          CPU 6: 5.1% user, 0.0% nice, 6.7% system, 17.3% interrupt, 71.0% idle
                          CPU 7: 4.3% user, 0.0% nice, 7.1% system, 11.0% interrupt, 77.6% idle

                          but it certainly looks clean

                          1 Reply Last reply Reply Quote 1
                          • w0wW
                            w0w @dsl.ottawa
                            last edited by

                            @dsl-ottawa said in New PPPoE backend, some feedback:

                            put PPPOE back, and ran the same tests.

                            • download 3gig from pfsense using speedtest-cli
                            • and normal 1.7-1.9 on my desktop

                            Is there any chance that some old limiters, rules, traffic shaping, or other configurations are still active?
                            If this is a clean install, I would suspect a bug in pfSense

                            1 Reply Last reply Reply Quote 0
                            • M
                              mr_nets @w0w
                              last edited by mr_nets

                              I make a script to monitor the queues of the parent interface IX2 and all data goes through the queue 0 on the RX side.

                              BTW my ISP use a VLAN (IX2 --> VLAN40 --> PPPoE) maybe it doesn't help to manage the queues correctly...

                              In 24 Hours
                              5a0e6b2d-e486-477e-a21d-885af26de75a-image.png

                              w0wW 1 Reply Last reply Reply Quote 0
                              • w0wW
                                w0w @mr_nets
                                last edited by w0w

                                @mr_nets
                                There may be only one queue on the network card, but packets can be distributed across CPU cores. Previously, due to netgraph, everything was handled by a single core. The deferred option slightly improved this, but it was a limited improvement—overall, performance was bottlenecked by a single CPU core. At least that limitation no longer exists now.

                                In theory, if the network queue never fills up, there shouldn't be any issues. I don't know what the theoretical maximum throughput of a PPPoE session is—maybe Netgate specialists have already tested this? I would expect something like 2.5–3.5 Gbps with multithreaded download.

                                1 Reply Last reply Reply Quote 0
                                • stephenw10S
                                  stephenw10 Netgate Administrator
                                  last edited by

                                  It still depends on the hardware but we were able to pass >8Gbps through a 6100 in raw testing: https://www.netgate.com/blog/optimizing-pppoe-performance-in-pfsense-software

                                  w0wW 1 Reply Last reply Reply Quote 1
                                  • w0wW
                                    w0w @stephenw10
                                    last edited by

                                    @stephenw10 said in New PPPoE backend, some feedback:

                                    It still depends on the hardware but we were able to pass >8Gbps through a 6100 in raw testing: https://www.netgate.com/blog/optimizing-pppoe-performance-in-pfsense-software

                                    Thanks, great read. Too bad I missed it, but I'm glad you brought it to me today. 🙂

                                    1 Reply Last reply Reply Quote 1
                                    • C
                                      claudio69
                                      last edited by

                                      I am testing pfSense 2.8 on my pppoe WAN setup that has been working for many years.
                                      I noticed that the weekly periodic reset I set on the pppoe WAN is not automatically performed on the weekend.
                                      If I try to change the setting manually the pppoe connection resets and reconnects but then it will no longer do so automatically

                                      1 Reply Last reply Reply Quote 0
                                      • stephenw10S
                                        stephenw10 Netgate Administrator
                                        last edited by

                                        You're setting that in the PPP settings on the interface?

                                        Interesting. That's an advanced setting that only applies to mpd5. It's hidden on the ppp interfaces page but looks like it's not hidden from the assigned interface and probably should be.

                                        C 1 Reply Last reply Reply Quote 0
                                        • C
                                          claudio69 @stephenw10
                                          last edited by

                                          @stephenw10

                                          I use this setting on a pppoe connection with variable ip.
                                          I need it to avoid the connection being reset by the ISP since I do not have a fixed IP.
                                          On pfSense from 2.5 to 2.7.2 it worked on all versions.Schermata del 2025-04-21 15-10-18.png

                                          1 Reply Last reply Reply Quote 0
                                          • stephenw10S
                                            stephenw10 Netgate Administrator
                                            last edited by

                                            I assume you have enabled if_pppoe?

                                            C 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post
                                            Copyright 2025 Rubicon Communications LLC (Netgate). All rights reserved.