Mirai
is an SDK for highly optimized on-device LLM inference.
import Mirai
let engine = MiraiEngine.shared
let model = ModelsRegistry.llama_3_2_1b_instruct
guard let files = try await engine.storage.download(model: model) else {
return
}
- You can optionally select a configuration for your specific use case from the following list:
general
,chat
,summarization
,classification
. - Choosing the appropriate configuration significantly increases inference speed by utilizing specialized optimization techniques.
let session = try await engine.session(model: model, modelFiles: files)
try await session.updateConfiguration(.forType(.general))
try await session.load()
If you want to get an answer for a specific prompt, use the text
input:
try await session.run(input: .text("Ultimate Question of Life, the Universe, and Everything")) { result in
if result.finished {
print(result.text)
}
}
You can also use a list of messages
as an input:
let messages = [
Message(text: "Hi", role: .user),
Message(text: "How can I help you?", role: .assistant),
Message(text: "Tell me a story", role: .user)
]
try await session.run(input: .messages(messages)) { result in
if result.finished {
print(result.text)
}
}
If you have any questions, just drop us a message at [email protected].