Usertransfer in HA-Modus



  • Hi there,

    i found an old entry and a feature request but no current information what about Usertransfer in HA-Modus. Means: Users are Loged in on Master-Server and also the Slave one knows about the login (and Preferences).
    I thought the feature was planed for the 2.4-Release :)

    Does anybody know the current process?

    Greets,

    MasterD


  • Rebel Alliance

    hello,
    high avaliability is not functionnal on the captive portal. At all. see https://redmine.pfsense.org/issues/97

    I am currently working on an implementation, ( https://github.com/Augustin-FL/pfsense/tree/captiveportal-db-sync ), but it will probably take me months to get something working

    the thing that was supposed to be added in 2.4.4 was in use/expired vouchers synchronization. the idea was not to provide HA but rather to provide a way to have two separate captive portals that would use the same list of vouchers. ( see https://redmine.pfsense.org/issues/7972 for details ). i can confirm that the backward sync is not working correctly in 2.4.4, as explained by the last comment of this issue


  • Rebel Alliance

    hi,
    I finally made a patch for it !

    I am now looking for testers (even if you are not using High Availability). Could anyone install this patch on a development Server (2.5.0) and give me some feedback?

    Here is how to install it :

    1. Install the patch package
    2. Create a new patch. In "URL/Commit ID", enter https://patch-diff.githubusercontent.com/raw/pfsense/pfsense/pull/4150.diff . Let the default settings in the "Patch Application Behavior" section (Path Strip Count : 2, etc...)
    3. Fetch and apply the new patch. After installing, reboot your pfSense.
    4. After installing the patch, if you wish to use High Availability for captive portal :
      • Configure High Availability normally using System->High Avail. Sync menu.
      • Configure XMLRPC sync on the primary node only, as it would be done for a normal configuration
      • on the secondary node, please go to Services->Captive Portal->(your zone)->High Availability and configure backward synchronization.

    How it works / Behavior

    When using HA,

    • In normal situation (both nodes UP), captive portal users and vouchers are synchronized between nodes.
    • If the primary node become unreachable, secondary node become master and continues to run the captive portal
    • If the primary node switch back from backup to master, it tries to refresh connected users from the secondary (and now backup) node.
    • If the secondary node leave then re-join the cluster, users will NOT be synchronized on the backup node. Users have to be manually synchronized from Captive Portal->Your CP zone->High Availability in such situation.

    What this patch is NOT / Limitations

    • This patch aims to sync connected users, and in-use/expired vouchers. Allowed IP addresses/hostnames/MACs synchronization are out of scope.
    • This patch is designed to handle a failure from the primary node, not from the secondary one. Because of the very way HA is implemented on pfSense, a failure on the secondary node would have some bad effects for the cluster. In the case of the captive portal, the effects would be some slowness when performing an user (dis)connection.
      This issue is not specific to captive portal, and is due to how how XMLRPC sync works in pfSense.
      • The workaround to this issue is to manually un-check Captive Portal in HA settings when secondary node leaves the cluster.
    • RADIUS accounting also works fine with HA, but per-user data consumption is not synchronized between nodes.

    Developer notes / technical info

    • This patch implement a new XMLRPC endpoint, pfsense.captive_portal_sync. It was necessary to implement this endpoint because of bi-directional synchronization (using pfsense.restore_config_section is causing many problems, such as triggering a DHCP server restart every time an user get connected)

    Please don't hesitate to comment if you have questions/feedback to share ! ☺


Log in to reply