Marziel OS v1.0.1

Private AI Operating System โ€” runs entirely on your hardware.

Install

pip install marziel==1.0.1
marziel serve

v1.0.1 โ€” Marziel OS

AI Kernel

  • Persistent event loop with autonomous decision-making
  • 3-Tier Memory: Working, Long-Term (AES-256), Episodic
  • Process Manager: Unix-like ps/top/kill
  • Task Scheduler for recurring tasks

MarzielFlow โ€” 5-Component Adaptive Inference

  • Speculative Decoding with automatic fallback
  • T/Z Distribution Quantization (Normal/Student-t/Beta)
  • Adaptive Bit-Width: 2.88-bit avg across 32 layers
  • Fuzzy Logic Controller (Mamdani-style)
  • Attention Sink Cache: 75% memory savings

TurboQuant

PolarQuant + QJL 3-bit KV cache โ€” 3x memory reduction.

Model Formats

Format Size Platform
GGUF Q4_K_M 4.8 GB NVIDIA GPU, CPU
MLX 4-bit 4.5 GB Apple Silicon
Safetensors 16 GB Full precision

GGUF Usage

from llama_cpp import Llama
model = Llama.from_pretrained(
    repo_id="efops/marziel-8b-custom",
    filename="marziel-v6-Q4_K_M.gguf",
    n_gpu_layers=-1, n_ctx=4096,
)
output = model.create_chat_completion(
    messages=[{"role": "user", "content": "Hello!"}]
)

MLX Usage

pip install mlx-lm
mlx_lm.generate --model efops/marziel-8b-custom-MLX --prompt "Hello!"

OS API

GET  /os/status    โ€” Kernel status
GET  /os/memory    โ€” Memory tiers
GET  /os/ps        โ€” Process list
GET  /os/top       โ€” Resource monitor
POST /os/recall    โ€” Memory recall
POST /os/remember  โ€” Store memory
POST /os/schedule  โ€” Schedule tasks
POST /os/kill/:pid โ€” Kill process

Performance

  • 52.9 tok/s on NVIDIA RTX A5000
  • 75% KV cache memory savings
  • 2.88-bit avg quantization

Links

MIT License โ€” Built by Efe (Efkan Isazade)

Downloads last month
2,544
MLX
Hardware compatibility
Log In to add your hardware

4-bit

GGUF
Model size
8B params
Architecture
mistral3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for efops/marziel-8b-custom

Quantized
(69)
this model