I have used mlx_lm.lora to fine tune a mistral-7b-v0.3-4bit model with my data. I fused the mistral model with my adapters and upload the fused model to my directory on huggingface. I was able to use mlx_lm.generate to use the fused model in Terminal. However, I don't know how to load the model in Swift. I've used
Imports
import SwiftUI
import MLX
import MLXLMCommon
import MLXLLM
        let modelFactory = LLMModelFactory.shared
        let configuration = ModelConfiguration(
            id: "pharmpk/pk-mistral-7b-v0.3-4bit"
        )
        
// Load the model off the main actor, then assign on the main actor
        let loaded = try await modelFactory.loadContainer(configuration: configuration)
        { progress in
            print("Downloading progress: \(progress.fractionCompleted * 100)%")
        }
        await MainActor.run {
            self.model = loaded
        }
I'm getting an error
runModel error: downloadError("A server with the specified hostname could not be found.")
Any suggestions?
Thanks, David
PS, I can load the model from the app bundle // directory: Bundle.main.resourceURL! but it's too big to upload for Testflight