Skip to content

Latest commit

 

History

History
162 lines (133 loc) · 4.53 KB

README.md

File metadata and controls

162 lines (133 loc) · 4.53 KB

Swollama

swollama-logo-small

Documentation License

A comprehensive, protocol-oriented Swift client for the Ollama API. This package provides a type-safe way to interact with Ollama's machine learning models, supporting all API endpoints with native Swift concurrency.

Table of Contents

Features

  • ✨ Full Ollama API coverage
  • 🔄 Native async/await and AsyncSequence support
  • 🛡️ Type-safe API with comprehensive error handling
  • 🔒 Thread-safe implementation using Swift actors
  • 🔄 Automatic retry logic for failed requests
  • 📦 Zero external dependencies

Requirements

  • macOS 14+
  • Xcode 15.0+
  • Swift 5.9+
  • Ollama installed and running locally or on a remote server

Installation

Swift Package Manager

Add Swollama to your Swift package dependencies in Package.swift:

dependencies: [
    .package(url: "https://github.com/marcusziade/Swollama.git", from: "1.0.0")
]

Or add it through Xcode:

  1. File > Add Package Dependencies
  2. Enter the repository URL: https://github.com/marcusziade/Swollama.git

Quick Start

import Swollama

// Initialize client
let client = OllamaClient()

// List available models
let models = try await client.listModels()
for model in models {
    print(model.name)
}

// Start a chat
guard let model = OllamaModelName.parse("llama3.2") else {
    throw CLIError.invalidArgument("Invalid model name format")
}
let responses = try await client.chat(
    messages: [
        ChatMessage(role: .user, content: "Hello! How are you?")
    ],
    model: model
)
for try await response in responses {
    print(response.message.content, terminator: "")
}

CLI Usage

Stream a chat response:

swollama chat llama3.2

CleanShot 2024-10-27 at 15 12 39

Generate text with specific parameters:

swollama generate codellama

Pull a new model:

swollama pull llama3.2

CleanShot 2024-10-27 at 15 19 34

List all available models:

swollama list

CleanShot 2024-10-27 at 15 24 28@2x

Show model information:

swollama show llama3.2

Copy a model:

swollama copy llama3.2 my-llama3.2

Delete a model:

swollama delete my-llama3.2

List running models:

swollama ps

Documentation

For complete API documentation, usage examples, and best practices, visit the Documentation.

Examples

Chat Completion

let client = OllamaClient()
let responses = try await client.chat(
    messages: [
        .init(role: .system, content: "You are a helpful assistant"),
        .init(role: .user, content: "Write a haiku about Swift")
    ],
    model: .init("llama3.2")!
)

for try await response in responses {
    print(response.message.content)
}

Generate Text

let client = OllamaClient()
let responses = try await client.generate(
    prompt: "Explain quantum computing",
    model: .init("llama3.2")!
)

for try await response in responses {
    print(response.content)
}

Contributing

Contributions are welcome! Please feel free to submit a Pull Request. For major changes, please open an issue first to discuss what you would like to change.

Contact

If you have any questions, feedback, or run into issues, please open an issue on the GitHub repository.