Skip to content
View Goekdeniz-Guelmez's full-sized avatar
🤨
AGI in 2026?
🤨
AGI in 2026?

Block or report Goekdeniz-Guelmez

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Goekdeniz-Guelmez/README.md

Gökdeniz Gülmez

ML Researcher · Open-Source Engineer · Apple Silicon ML Ecosystem

GitHub followers arXiv Sponsor

Stuttgart, Germany — building the infrastructure that makes local AI actually work.


I build the tools that let researchers and engineers run, train, and understand large language models on their own hardware — specifically Apple Silicon. My work spans core MLX contributions, independent research published on arXiv, and open-source projects used by thousands.

If my work has saved you GPU bills, unlocked fine-tuning on your Mac, or ended up in your pipeline — consider sponsoring. Everything I build is free, maintained in my spare time, and driven by the belief that capable AI tooling shouldn't require a data center.


Research

Paper Description Year
JOSIEfied Qwen3.5 Gabliterated Models Newest iteration of the JOSIEfied model family — now with vision support 2026
DynaMoE Dynamic, adaptive Mixture-of-Experts LLM architecture 2026
JOSIE Models World's first fully fine-tuned model family trained entirely on Apple Silicon 2026
JOSIEfied Qwen3 Abliterated Models Reached #1 globally on uncensored benchmarks 2025
Gabliteration Automated Gabliteration for any Transformers-compatible LLM 2025

Open-Source Projects

Train large language models natively on Apple Silicon

LoRA, QLoRA, and full-precision fine-tuning for LLMs — built on MLX. Supports Preference Optimization, RLHF, RL with and custom reward functions. The go-to fine-tuning toolkit for anyone running Apple Silicon.

  • Full-precision, LoRA, and QLoRA training modes
  • 12+ training methods and algorythms
  • WandB integration for training metrics
  • Multiple optimizer support including Muon
  • Example notebooks

Abliterate and research the inner workings of LLMs

Interpretability and intervention tooling for language models in MLX. Understand what's happening inside your model, not just what comes out.


Your own Google's NotebookLM — fully local, fully private

PDF-grounded audio generation (podcasts, summaries, and many more with up to 6 speakers) running entirely on-device. No API keys, no cloud, no data leaving your machine. As well as a compation native App in Local NotebookLM-App.


Fine-tune embedding models natively on Apple Silicon

Train and adapt embedding models for retrieval, search, and semantic tasks — directly in MLX.


Benchmark LLMs on Apple MLX framework knowledge and coding tasks.

MLX Benchmark is the first comprehensive CLI tool and dataset that measures how well large language models understand, write, and debug code for Apple's MLX machine learning framework — covering everything from core array operations to LoRA fine-tuning with mlx-lm, mlx-vlm, and mlx-embeddings.


📐 MLX-KAN

Kolmogorov-Arnold Networks in MLX

Native MLX implementation of KANs — a fundamentally different alternative to MLPs, implemented cleanly for Apple Silicon.


Automated abliteration for any Transformers LLM

Companion code to arXiv:2412.06527. Remove refusal directions from any model supported in Hugging Face Transformers.


A family of fine-tuned models that reached #1 globally on relevant benchmarks (arXiv:2512.18901).


Contributions to the MLX Ecosystem

I'm an officially acknowledged contributor to the core MLX stack:

Architectures I've added to MLX / MLX-LM:

Expand full list
Model Organization
Mamba v1 & v2 State Space
MiniCPM & MiniCPM3 OpenBMB
Helium Kyutai
GLM, GLM4, GLM5 Z.ai & THUKEG
dots.llm1 Rednote
Ernie4.5 MoE Baidu
Bailing MoE & Bailing MoE Linear (Ling-family) inclusionAI
Granite MoE IBM
LongCat Meituan
Nemotron H NVIDIA
Apertus Swiss-AI
OLMoE & OLMo 3 AllenAI
Jamba AI21 Labs
And many more...

Training features I've contributed:

  • Full weight fine-tuning support in mlx-lm
  • Muon optimizer (both mlx and mlx-lm)
  • WandB metric reporting during training
  • Multiple optimizer choices for training runs
  • ReLU² activation function in core MLX

Currently Building

J.O.S.I.E.-Home

A fully local, real-time, full-duplex multimodal assistant for smart home control

A discrete diffusion language model with a custom tokenizer (ChatML-style format, hardcoded vocabulary covering rooms, devices, properties, and continuous value bins) that can autonomously control and manage a complete smart home environment — sensors, cameras, LEDs, and more. Fully offline. No cloud dependency.

Current focus: training data generation strategies and model architecture validation.

Josie-Linear

A new Linear Dynamic Mixture-of-Experts LLM architecture — currently in development.


Support This Work

Everything above — the MLX contributions, the research, the open-source tools — is built in my spare time. If any of it has been useful to you or your team, sponsoring directly funds continued development, faster bug fixes, and new features.

Sponsor Gökdeniz


GitHub Activity

GitHub activity graph


Stuttgart, Germany · ML Researcher & Engineer · Apple Silicon AI Tooling

Pinned Loading

  1. mlx-lm-lora mlx-lm-lora Public

    Train Large Language Models on MLX.

    Python 291 42

  2. mlx-lm-lora-example-notebooks mlx-lm-lora-example-notebooks Public

    this repo has all official MLX-LM-LoRA example notebooks for training on Apple Silicon

    Jupyter Notebook 30 5

  3. mlx-kan mlx-kan Public

    KAN (Kolmogorov–Arnold Networks) in the MLX framework for Apple Silicon

    Python 31 4

  4. mlx-embeddings-lora mlx-embeddings-lora Public

    Train Embedding Models on MLX.

    Python 16 2

  5. Local-NotebookLM Local-NotebookLM Public

    Googles NotebookLM but local

    Python 880 112

  6. gabliteration gabliteration Public

    Automated hyperparameter search for optimal Gabliteration configurations on large language models

    Python 49 8