Unfortunately a 13B model probably isn't going to cut it. Perchance uses a popular open source 70B Llama-based model (you'll come across it's name almost immediately if you look at top model lists, but any of the top models will work fine - and you should use the recommended parameters in the HuggingFace repo). If you can't run a 70B models, then I'd recommend these two places to find a 30B/20B/13B model to suit your specific use case, depending on your GPU size:
- https://www.reddit.com/r/LocalLLaMA/top/?t=month (focuses mainly on intelligence - e.g. assistants)
- https://www.reddit.com/r/SillyTavernAI/top/?t=month (focuses on entertainment - e.g. story/rp/text-adventure)
This community is not well-suited to helping you get it set up, but the above two communities have lots of info.