Hmm, after testing the raw Phi 2 within LM Studio instead of the examples provided by HuggingFace candle, I think it's actually pretty decent after all.
Specifically I got phi 2 Q4_K_S gguf working from TheBloke.
Can't get that model running with candle since it can't seem to load the model weight.