• 5 Posts
  • 16 Comments
Joined 1 year ago
cake
Cake day: June 17th, 2023

help-circle


  • Hello @theit8514

    You are actually spot on ^^

    I did look in my exports file which was like so :/mnt/DiskArray 192.168.0.16(rw) 192.168.0.65(rw)

    I added a localhost line in case: /mnt/DiskArray 127.0.0.1(rw) 192.168.0.16(rw) 192.168.0.65(rw)

    It didn’t solve the problem. I went to investigate with the mount command:

    • Will mount on 192.168.0.65: mount -t nfs 192.168.0.55:/mnt/DiskArray/mystuff/ /tmp/test

    • Will NOT mount on 192.168.0.55 (NAS): mount -t nfs 192.168.0.55:/mnt/DiskArray/mystuff/ /tmp/test

    • Will mount on 192.168.0.55 (NAS): mount -t nfs 127.0.0.1:/mnt/DiskArray/mystuff/ /tmp/test

    The mount -t nfs 192.168.0.55 is the one that the cluster does actually. So i either need to find a way for it to use 127.0.0.1 on the NAS machine, or use a hostname that might be better to resolve

    EDIT:

    I was acutally WAY simpler.

    I just added 192.168.0.55 to my /etc/exports file. It works fine now ^^

    Thanks a lot for your help@theit8514@lemmy.world !







  • Hello ! Thanks for your response!

    Yes RAID is used as availability of my data here, with or without longhorn, there wouldn’t be much difference there (especially since i only use one specific node)

    And you would be right, since the other nodes are unscheduled, it will be available only on my “storage node” so if this one goes down my storage goes down.

    That’s why i might be overkill with longhorn, but there are functions to restore and backup to s3 for exemple that i would need to setup i guess