Negative partition size in manually partitioning of 2.4.4-snapshot
abcnew last edited by abcnew
Today I used 20180827-0743 image of 2.4.4-snapshot to do a fresh install on a bare metal machine with a 150G SSD. When I am manually partitioning with root, usr, home, tmp and var partitions, the size of ada0p3(sometime ada0p4) would automatically turn into negative(displaying 0.1T outside but displaying negative size in "modify"ing it) when finished allocating all the partitions' sizes and the installing program cannot write data into that partition then returned to select the keymap(like en-US).
Auto UFS mode had no problem and runs well.
I tried 2.4.3 release' iso in virtualbox. No such problem in manually partitioning. And it seems no way to let virtualbox to boot from a UEFI FreeBSD flash drive. So I am not able to install 2.4.4 snapshot in virtualbox.
I hope I can get a ISO image of 2.4.4 snapshot so I can try it in virtualbox or someone can fix this negative size issue of 2.4.4 snapshot.
You can get a 2.4.4 ISO right now. Download the memstick, decompress it, then rename
.iso-- It's a hybrid image that works both ways.
I haven't seen any negative size issue here but I also do not have that large of a drive spare to test at the moment.
Why manually partition though? There is very little benefit to doing that on a firewall.
abcnew last edited by
Thank you, @jimp
Using the image of 2.4.4-20180827-0743 in virtualbox, I found nothing wrong in manually ufs partitioning. I installed a win10 in that SSD before. I only removed all the partitions on it before intalling this pfsense snapshot. I think these things may cause the manually partitioning has negative size on ada0p3 or ada0p4.
By the way, I thought by setting the sizes of "/", "/home", and swap partitions, there would be always some spare space to let the SSD live longer, especially when running squid and snort. This knowledge is from the hardware board of this forum.
That is most likely outdated advice.