Netgate Discussion Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Search
    • Register
    • Login

    x11sdv-8c-tp8f (xeon-d 2146NT) compatibility?

    Scheduled Pinned Locked Moved Hardware
    11 Posts 4 Posters 1.3k Views
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • stephenw10S
      stephenw10 Netgate Administrator
      last edited by

      Not sure I've seen anyone try it. I would say it's massively overpowered for most pfSense installs.
      You can check the list of supported hardware in the ix driver which would be required for the 10GbE NICs.
      https://github.com/pfsense/FreeBSD-src/blob/e6b497fa0a30a3ae23e1cb0825bc2d7b7cc3635c/sys/dev/ixgbe/if_ix.c#L59

      I would expect the i350 to work no problems though.

      There may be other components that require support.

      Steve

      1 Reply Last reply Reply Quote 0
      • Q
        q54e3w
        last edited by

        just following up for others... all interfaces powered up with no fuss with 2.4.4.

        igb0: <Intel(R) PRO/1000 Network Connection, Version - 2.5.3-k> port 0x9060-0x907f mem 0xe0c60000-0xe0c7ffff,0xe0c8c000-0xe0c8ffff at device 0.0 numa-domain 0 on pci8
        igb0: Using MSIX interrupts with 9 vectors
        igb0: Ethernet address: ac:1f:6b:73:87:e0
        igb0: Bound queue 0 to cpu 0
        igb0: Bound queue 1 to cpu 1
        igb0: Bound queue 2 to cpu 2
        igb0: Bound queue 3 to cpu 3
        igb0: Bound queue 4 to cpu 4
        igb0: Bound queue 5 to cpu 5
        igb0: Bound queue 6 to cpu 6
        igb0: Bound queue 7 to cpu 7
        igb0: netmap queues/slots: TX 8/1024, RX 8/1024
        igb1: <Intel(R) PRO/1000 Network Connection, Version - 2.5.3-k> port 0x9040-0x905f mem 0xe0c40000-0xe0c5ffff,0xe0c88000-0xe0c8bfff at device 0.1 numa-domain 0 on pci8
        igb1: Using MSIX interrupts with 9 vectors
        igb1: Ethernet address: ac:1f:6b:73:87:e1
        igb1: Bound queue 0 to cpu 8
        igb1: Bound queue 1 to cpu 9
        igb1: Bound queue 2 to cpu 10
        igb1: Bound queue 3 to cpu 11
        igb1: Bound queue 4 to cpu 12
        igb1: Bound queue 5 to cpu 13
        igb1: Bound queue 6 to cpu 14
        igb1: Bound queue 7 to cpu 15
        igb1: netmap queues/slots: TX 8/1024, RX 8/1024
        igb2: <Intel(R) PRO/1000 Network Connection, Version - 2.5.3-k> port 0x9020-0x903f mem 0xe0c20000-0xe0c3ffff,0xe0c84000-0xe0c87fff at device 0.2 numa-domain 0 on pci8
        igb2: Using MSIX interrupts with 9 vectors
        igb2: Ethernet address: ac:1f:6b:73:87:e2
        igb2: Bound queue 0 to cpu 0
        igb2: Bound queue 1 to cpu 1
        igb2: Bound queue 2 to cpu 2
        igb2: Bound queue 3 to cpu 3
        igb2: Bound queue 4 to cpu 4
        igb2: Bound queue 5 to cpu 5
        igb2: Bound queue 6 to cpu 6
        igb2: Bound queue 7 to cpu 7
        igb2: netmap queues/slots: TX 8/1024, RX 8/1024
        igb3: <Intel(R) PRO/1000 Network Connection, Version - 2.5.3-k> port 0x9000-0x901f mem 0xe0c00000-0xe0c1ffff,0xe0c80000-0xe0c83fff at device 0.3 numa-domain 0 on pci8
        igb3: Using MSIX interrupts with 9 vectors
        igb3: Ethernet address: ac:1f:6b:73:87:e3
        igb3: Bound queue 0 to cpu 8
        igb3: Bound queue 1 to cpu 9
        igb3: Bound queue 2 to cpu 10
        igb3: Bound queue 3 to cpu 11
        igb3: Bound queue 4 to cpu 12
        igb3: Bound queue 5 to cpu 13
        igb3: Bound queue 6 to cpu 14
        igb3: Bound queue 7 to cpu 15
        igb3: netmap queues/slots: TX 8/1024, RX 8/1024
        
        ixl0: <Intel(R) Ethernet Connection 700 Series PF Driver, Version - 1.9.9-k> mem 0xfa000000-0xfaffffff,0xfb018000-0xfb01ffff at device 0.0 numa-domain 0 on pci14
        ixl0: using 1024 tx descriptors and 1024 rx descriptors
        ixl0: fw 3.1.57069 api 1.5 nvm 3.33 etid 80001006 oem 1.262.0
        ixl0: PF-ID[0]: VFs 32, MSIX 129, VF MSIX 5, QPs 384, MDIO shared
        ixl0: Using MSIX interrupts with 9 vectors
        ixl0: Allocating 8 queues for PF LAN VSI; 8 queues active
        ixl0: Ethernet address: ac:1f:6b:73:89:70
        ixl0: SR-IOV ready
        queues is 0xfffffe00017f4000
        ixl0: netmap queues/slots: TX 8/1024, RX 8/1024
        ixl1: <Intel(R) Ethernet Connection 700 Series PF Driver, Version - 1.9.9-k> mem 0xf9000000-0xf9ffffff,0xfb010000-0xfb017fff at device 0.1 numa-domain 0 on pci14
        ixl1: using 1024 tx descriptors and 1024 rx descriptors
        ixl1: fw 3.1.57069 api 1.5 nvm 3.33 etid 80001006 oem 1.262.0
        ixl1: PF-ID[1]: VFs 32, MSIX 129, VF MSIX 5, QPs 384, MDIO shared
        ixl1: Using MSIX interrupts with 9 vectors
        ixl1: Allocating 8 queues for PF LAN VSI; 8 queues active
        ixl1: Ethernet address: ac:1f:6b:73:89:71
        ixl1: SR-IOV ready
        queues is 0xfffffe000192c000
        ixl1: netmap queues/slots: TX 8/1024, RX 8/1024
        ixl2: <Intel(R) Ethernet Connection 700 Series PF Driver, Version - 1.9.9-k> mem 0xf8000000-0xf8ffffff,0xfb008000-0xfb00ffff at device 0.2 numa-domain 0 on pci14
        ixl2: using 1024 tx descriptors and 1024 rx descriptors
        ixl2: fw 3.1.57069 api 1.5 nvm 3.33 etid 80001006 oem 1.262.0
        ixl2: PF-ID[2]: VFs 32, MSIX 129, VF MSIX 5, QPs 384, I2C
        ixl2: Using MSIX interrupts with 9 vectors
        ixl2: Allocating 8 queues for PF LAN VSI; 8 queues active
        ixl2: Ethernet address: ac:1f:6b:73:89:72
        ixl2: SR-IOV ready
        queues is 0xfffffe0001a82000
        ixl2: netmap queues/slots: TX 8/1024, RX 8/1024
        ixl3: <Intel(R) Ethernet Connection 700 Series PF Driver, Version - 1.9.9-k> mem 0xf7000000-0xf7ffffff,0xfb000000-0xfb007fff at device 0.3 numa-domain 0 on pci14
        ixl3: using 1024 tx descriptors and 1024 rx descriptors
        ixl3: fw 3.1.57069 api 1.5 nvm 3.33 etid 80001006 oem 1.262.0
        ixl3: PF-ID[3]: VFs 32, MSIX 129, VF MSIX 5, QPs 384, I2C
        ixl3: Using MSIX interrupts with 9 vectors
        ixl3: Allocating 8 queues for PF LAN VSI; 8 queues active
        ixl3: Ethernet address: ac:1f:6b:73:89:73
        ixl3: SR-IOV ready
        queues is 0xfffffe0001bc6000
        ixl3: netmap queues/slots: TX 8/1024, RX 8/1024
        
        CPU: Intel(R) Xeon(R) D-2146NT CPU @ 2.30GHz (2294.68-MHz K8-class CPU)
          Origin="GenuineIntel"  Id=0x50654  Family=0x6  Model=0x55  Stepping=4
          Features=0xbfebfbff<FPU,VME,DE,PSE,TSC,MSR,PAE,MCE,CX8,APIC,SEP,MTRR,PGE,MCA,CMOV,PAT,PSE36,CLFLUSH,DTS,ACPI,MMX,FXSR,SSE,SSE2,SS,HTT,TM,PBE>
          Features2=0x7ffefbff<SSE3,PCLMULQDQ,DTES64,MON,DS_CPL,VMX,SMX,EST,TM2,SSSE3,SDBG,FMA,CX16,xTPR,PDCM,PCID,DCA,SSE4.1,SSE4.2,x2APIC,MOVBE,POPCNT,TSCDLT,AESNI,XSAVE,OSXSAVE,AVX,F16C,RDRAND>
          AMD Features=0x2c100800<SYSCALL,NX,Page1GB,RDTSCP,LM>
          AMD Features2=0x121<LAHF,ABM,Prefetch>
          Structured Extended Features=0xd39ffffb<FSGSBASE,TSCADJ,BMI1,HLE,AVX2,FDPEXC,SMEP,BMI2,ERMS,INVPCID,RTM,PQM,NFPUSG,MPX,PQE,AVX512F,AVX512DQ,RDSEED,ADX,SMAP,CLFLUSHOPT,CLWB,PROCTRACE,AVX512CD,AVX512BW,AVX512VL>
          Structured Extended Features2=0x8<PKU>
          Structured Extended Features3=0x9c000000<IBPB,STIBP,SSBD>
          XSAVE Features=0xf<XSAVEOPT,XSAVEC,XINUSE,XSAVES>
          VT-x: PAT,HLT,MTF,PAUSE,EPT,UG,VPID,VID,PostIntr
          TSC: P-state invariant, performance statistics
        
        
        1 Reply Last reply Reply Quote 1
        • planedropP
          planedrop
          last edited by

          I know this is quite the old thread, but instead of starting a new one, figured I'd ask a few things here first.

          @q54e3w did you end up getting everything working on the 2146NT? I have a need for a proper 10 gigabit real world performance WAN/NAT firewall and am considering this CPU platform for it since it's newer than the 1541's hardware and has QAT built in. Did you manage to get QAT working as well?

          Also debating if I just keep using my Netgate 6100 for now and wait for the possibility of something higher end than the 8200 to come out (that is newer than the 1541).

          Q 1 Reply Last reply Reply Quote 0
          • Q
            q54e3w @planedrop
            last edited by q54e3w

            @planedrop Appears to work fine.

            97331bc4-2bfa-4532-bcdd-5bfd7e552d1a-image.png

            planedropP Dobby_D 2 Replies Last reply Reply Quote 3
            • planedropP
              planedrop @q54e3w
              last edited by

              @q54e3w OK this is fantastic news, thanks a ton! Likely going with the same platform for my next firewall. Unless Netgate releases something more up-to-date than the 1541 before I pull the trigger on this lol

              1 Reply Last reply Reply Quote 0
              • Dobby_D
                Dobby_ @q54e3w
                last edited by

                @q54e3w

                Thanks from me too, for showing up the informations!

                It would also my next platform for pfSense (pfSense+) it offers all for internal installing,
                it comes with 3 M.2/miniPCIe slots for modem SSD and WiFi for internal installing.

                • Intel TurboBoost
                • Intel Hyper-Threading
                • Intel Quick Assist
                • AES-NI

                But one thing is irritating me a little bit, in your image is shown that there will be only 8 Cores and max. 2.3GHz
                at the Intel site it is shown something like 8 cores and 16 threads and max. 3.0 GHz. So my question is the following,
                have enabled Intel Hyper-threading and TurboBoost in the BIOS?

                #~. @Dobby

                Turris Omnia - 4 Ports - 2 GB RAM / TurrisOS 7 Release (Btrfs)
                PC Engines APU4D4 - 4 Ports - 4 GB RAM / pfSense CE 2.7.2 Release (ZFS)
                PC Engines APU6B4 - 4 Ports - 4 GB RAM / pfSense+ (Plus) 24.03_1 Release (ZFS)

                planedropP 1 Reply Last reply Reply Quote 0
                • planedropP
                  planedrop @Dobby_
                  last edited by

                  @Dobby_ they may have SMT disabled, pfSense does support it though.

                  1 Reply Last reply Reply Quote 1
                  • stephenw10S
                    stephenw10 Netgate Administrator
                    last edited by

                    Yes, if it was enabled it would show on the dashboard there.

                    1 Reply Last reply Reply Quote 1
                    • Q
                      q54e3w
                      last edited by

                      yes, Hyperthreading is disabled on this box hence the 8/8 rather than 8/16 you should expect.
                      Its also running a XL710L dual port card for multi-gig 2.5/5gbs support. It was originally supporting a 2.5g connection from a cable modem enabling 1.2gbps service.
                      The systems been 100% reliable running 4-5 load balanced OpenVPN connections and several wireguard tunnels too.
                      I do think its massively overpowered though.

                      planedropP 1 Reply Last reply Reply Quote 1
                      • planedropP
                        planedrop @q54e3w
                        last edited by

                        @q54e3w Thanks for all the info here, glad to hear it's working well and has been stable. I'm only considering it since I have an 8 gigabit WAN connection now.

                        1 Reply Last reply Reply Quote 0
                        • First post
                          Last post
                        Copyright 2025 Rubicon Communications LLC (Netgate). All rights reserved.