| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Tue, 09 Dec 2025 01:31:53 GMT
access-control-allow-origin: *
etag: W/"69377c09-19e6c1"
expires: Mon, 29 Dec 2025 02:06:46 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 3EBE:2F7ECD:829433:92B7CC:6951DFDD
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 01:56:46 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210025-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1766973407.623860,VS0,VE272
vary: Accept-Encoding
x-fastly-request-id: b9f45b9123a36253ffa48f5a31aeb82ef44f18b5
content-length: 1044626
MLC LLM | Home
MLC LLM: Universal LLM Deployment Engine With ML Compilation
Overview
MLC LLM is a machine learning compiler and high-performance deployment engine for large language models. The mission of this project is to enable everyone to develop, optimize, and deploy AI models natively on everyone’s platforms.
MLC LLM compiles and runs code on MLCEngine – a unified high-performance LLM inference engine across the above platforms. MLCEngine provides OpenAI-compatible API available through REST server, python, javascript, iOS, Android, all backed by the same engine and compiler that we keep improving with the community.
Get Started
Please visit our documentation to get started with MLC LLM.