Downloading my fine tuned model from huggingface

I have used mlx_lm.lora to fine tune a mistral-7b-v0.3-4bit model with my data. I fused the mistral model with my adapters and upload the fused model to my directory on huggingface. I was able to use mlx_lm.generate to use the fused model in Terminal. However, I don't know how to load the model in Swift. I've used

Imports

import SwiftUI
import MLX
import MLXLMCommon
import MLXLLM
        let modelFactory = LLMModelFactory.shared
        let configuration = ModelConfiguration(
            id: "pharmpk/pk-mistral-7b-v0.3-4bit"
        )
        
// Load the model off the main actor, then assign on the main actor
        let loaded = try await modelFactory.loadContainer(configuration: configuration)
        { progress in
            print("Downloading progress: \(progress.fractionCompleted * 100)%")
        }
        await MainActor.run {
            self.model = loaded
        }

I'm getting an error

runModel error: downloadError("A server with the specified hostname could not be found.")

Any suggestions?

Thanks, David

PS, I can load the model from the app bundle // directory: Bundle.main.resourceURL! but it's too big to upload for Testflight

Downloading my fine tuned model from huggingface
 
 
Q