Load → Inspect → Test
Load a model
- Click Load (01) in the sidebar
- Click LOAD GGUF FILE (or LOAD SAFETENSORS FILE / LOAD MODEL FOLDER)
- Select your model file in the dialog
- Wait for header parsing (< 1 second)
Inspect the architecture
- Click Inspect (02) in the sidebar
- Explore the isometric 3D visualization — hover layers for details
- Review memory distribution, quantization breakdown, and runtime compatibility
- Check capability detection (reasoning, code, math, etc.)
For GGUF inference, llama.cpp tools must be installed. Go to Settings (07) → llama.cpp Tools → DOWNLOAD & INSTALL.
Download → Quantize
Download from HuggingFace
- Click Hub (04) in the sidebar
- Enter a repo ID (e.g.,
TheBloke/Mistral-7B-Instruct-v0.2-GGUF) - Click FETCH, then DOWNLOAD on a GGUF file
Explore a Dataset
Load from HuggingFace
- Click DataStudio (10) in the sidebar
- Switch to HUGGINGFACE source
- Enter a dataset repo ID (e.g.,
tatsu-lab/alpaca) - Click FETCH to see available files
- Click DOWNLOAD on a Parquet or JSON file
- Dataset auto-loads after download
Fine-Tune a Model
Select model and dataset
- Click Training (06) in the sidebar
- Browse for a model (GGUF or SafeTensors)
- Browse for a dataset (JSON, JSONL, CSV, or Parquet)
Choose method and preset
- Select a training method (LoRA, QLoRA, SFT, DPO, Full)
- Pick a VRAM preset (LOW VRAM ~4GB, BALANCED ~6GB, QUALITY ~12GB, MAX QUALITY ~24GB)
- Optionally target specific capabilities (reasoning, code, math, etc.)
Merge Two Models
Load parents
- Click M-DNA (08) in the sidebar
- In the Files tab, click + LOAD FILE twice to load 2 models
Configure merge
- Go to Settings tab, pick a preset (QUICK BLEND, SMOOTH MERGE, TASK TUNER, etc.)
- Or manually select a method and tune parameters
- Set output format (SafeTensors or GGUF) and output path