Selfhosted
A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.
Rules:
-
Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.
-
No spam posting.
-
Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.
-
Don't duplicate the full text of your blog or github here. Just post the link for folks to click.
-
Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).
-
No trolling.
Resources:
- selfh.st Newsletter and index of selfhosted software and apps
- awesome-selfhosted software
- awesome-sysadmin resources
- Self-Hosted Podcast from Jupiter Broadcasting
Any issues on the community? Report it using the report flag.
Questions? DM the mods!
view the rest of the comments
Ya I realize this isn’t a great way to go about storage but I already have the enclosure so I might as well use it for now. At some point down the line I will build something that will work better.
If I connect it using USB I am able to see each drive individually in Proxmox. I am unsure if it will be the same if I use eSATA. In the manual it says that the eSATA interface card needs to support Port Multiplier which I fear means the eSATA to SATA option may not work but I was hoping someone here may know more about that.
If I have to go the USB route and I am able to use each drive individually, would you recommend going with a ZFS pool or ceph?
I'd go with ZFS because balancing between the included disk in a Ceph pool will consume I/O and especially if one disk fails, that rebuild would take days if not weeks. If you want to try Ceph then don't include all the 8 drives initially and see how the performance is with the minimum 3 drives.
Thanks, I may hold off on ceph for now in that case