axiom-ios-ml

📁 fotescodev/ios-agent-skills 📅 5 days ago
3
总安装量
3
周安装量
#56894
全站排名
安装命令
npx skills add https://github.com/fotescodev/ios-agent-skills --skill axiom-ios-ml

Agent 安装分布

opencode 3
antigravity 3
claude-code 3
github-copilot 3
codex 3
kimi-cli 3

Skill 文档

iOS Machine Learning Router

You MUST use this skill for ANY on-device machine learning or speech-to-text work.

When to Use

Use this router when:

  • Converting PyTorch/TensorFlow models to CoreML
  • Deploying ML models on-device
  • Compressing models (quantization, palettization, pruning)
  • Working with large language models (LLMs)
  • Implementing KV-cache for transformers
  • Using MLTensor for model stitching
  • Building speech-to-text features
  • Transcribing audio (live or recorded)

Routing Logic

CoreML Work

Implementation patterns → /skill coreml

  • Model conversion workflow
  • MLTensor for model stitching
  • Stateful models with KV-cache
  • Multi-function models (adapters/LoRA)
  • Async prediction patterns
  • Compute unit selection

API reference → /skill coreml-ref

  • CoreML Tools Python API
  • MLModel lifecycle
  • MLTensor operations
  • MLComputeDevice availability
  • State management APIs
  • Performance reports

Diagnostics → /skill coreml-diag

  • Model won’t load
  • Slow inference
  • Memory issues
  • Compression accuracy loss
  • Compute unit problems

Speech Work

Implementation patterns → /skill speech

  • SpeechAnalyzer setup (iOS 26+)
  • SpeechTranscriber configuration
  • Live transcription
  • File transcription
  • Volatile vs finalized results
  • Model asset management

Decision Tree

User asks about on-device ML or speech
  ├─ Machine learning?
  │   ├─ Implementing/converting? → coreml
  │   ├─ Need API reference? → coreml-ref
  │   └─ Debugging issues? → coreml-diag
  └─ Speech-to-text?
      └─ Any speech work → speech

Critical Patterns

coreml:

  • Model conversion (PyTorch → CoreML)
  • Compression (palettization, quantization, pruning)
  • Stateful KV-cache for LLMs
  • Multi-function models for adapters
  • MLTensor for pipeline stitching
  • Async concurrent prediction

coreml-diag:

  • Load failures and caching
  • Inference performance issues
  • Memory pressure from models
  • Accuracy degradation from compression

speech:

  • SpeechAnalyzer + SpeechTranscriber setup
  • AssetInventory model management
  • Live transcription with volatile results
  • Audio format conversion

Example Invocations

User: “How do I convert a PyTorch model to CoreML?” → Invoke: /skill coreml

User: “Compress my model to fit on iPhone” → Invoke: /skill coreml

User: “Implement KV-cache for my language model” → Invoke: /skill coreml

User: “Model loads slowly on first launch” → Invoke: /skill coreml-diag

User: “My compressed model has bad accuracy” → Invoke: /skill coreml-diag

User: “Add live transcription to my app” → Invoke: /skill speech

User: “Transcribe audio files with SpeechAnalyzer” → Invoke: /skill speech

User: “What’s MLTensor and how do I use it?” → Invoke: /skill coreml-ref