| | Microserving LLM Engines (mlc.ai) |
|
1 point by homarp 31 days ago | past | 1 comment
|
| | LLM Microserving: a new RISC-style approach to design LLM serving API (mlc.ai) |
|
4 points by jinhongyii 37 days ago | past | 1 comment
|
| | Making AMD GPUs competitive for LLM inference (2023) (mlc.ai) |
|
313 points by plasticchris 52 days ago | past | 213 comments
|
| | Optimizing and Characterizing High-Throughput Low-Latency LLM Inference (mlc.ai) |
|
1 point by djhu9 4 months ago | past
|
| | High-Throughput Low-Latency LLM Serving with MLCEngine (mlc.ai) |
|
8 points by ruihangl 4 months ago | past
|
| | In-browser LLM inference engine with WebGPU and OpenAI API (mlc.ai) |
|
16 points by CharlieRuan 8 months ago | past | 4 comments
|
| | MLCEngine: Universal LLM Deployment to Both Cloud and Local Devices (mlc.ai) |
|
2 points by crowwork 8 months ago | past
|
| | Universal LLM Deployment Engine with ML Compilation (mlc.ai) |
|
17 points by ruihangl 8 months ago | past | 7 comments
|
| | MLC LLM: Universal Language Model Deployment Across Diverse Hardware and Apps (mlc.ai) |
|
1 point by georgehill on Dec 16, 2023 | past
|
| | Scaling LLama2-70B with Multiple Nvidia/AMD GPU (mlc.ai) |
|
13 points by junrushao1994 on Oct 20, 2023 | past | 6 comments
|
| | WebLLM: Llama2 in the Browser (mlc.ai) |
|
192 points by meiraleal on Aug 29, 2023 | past | 31 comments
|
| | GPU-Accelerated LLM on an Orange Pi (mlc.ai) |
|
214 points by tosh on Aug 15, 2023 | past | 80 comments
|
| | Making AMD GPUs competitive for LLM inference (mlc.ai) |
|
354 points by djoldman on Aug 9, 2023 | past | 132 comments
|
| | Run Llama2-70B in Web Browser with WebGPU Acceleration (mlc.ai) |
|
9 points by ruihangl on July 24, 2023 | past | 6 comments
|
| | Bringing Open Large Language Models to Consumer Devices (mlc.ai) |
|
31 points by hardmaru on May 23, 2023 | past
|
| | Running RedPajama and other open LLMs on phones, browsers and AMD/NV/Intel GPUs (mlc.ai) |
|
11 points by junrushao1994 on May 23, 2023 | past
|
| | Bringing Open Large Language Models to Consumer Devices (mlc.ai) |
|
11 points by shantanu_sharma on May 22, 2023 | past
|
| | Browser-based Stable Diffusion using WebGPU (mlc.ai) |
|
3 points by Eduard on May 6, 2023 | past
|
| | Bringing Hardware Accelerated Language Models to Consumer Devices (mlc.ai) |
|
1 point by crowwork on May 1, 2023 | past
|
| | MLC: Bringing Hardware Accelerated Language Models to Consumer Devices (mlc.ai) |
|
8 points by junrushao1994 on May 1, 2023 | past
|
| | What Is ML Compilation (mlc.ai) |
|
88 points by tosh on April 30, 2023 | past | 5 comments
|
| | Vicuna on iPhone (mlc.ai) |
|
90 points by tosh on April 30, 2023 | past | 15 comments
|
| | MLC LLM (mlc.ai) |
|
2 points by tosh on April 30, 2023 | past
|
| | MLC LLM – Large Language Models on iPhone GPU and Many More GPU Platforms (mlc.ai) |
|
2 points by crowwork on April 29, 2023 | past
|
| | Web LLM (mlc.ai) |
|
89 points by synergy20 on April 25, 2023 | past | 15 comments
|
| | Web LLM: Bringing large-language models and chat to web browsers (mlc.ai) |
|
3 points by yarapavan on April 17, 2023 | past
|
| | Web LLM – Vicuna (like LLaMA) locally inside the browser (mlc.ai) |
|
5 points by the_tli on April 15, 2023 | past
|
| | WebLLM (mlc.ai) |
|
6 points by marban on April 15, 2023 | past
|
| | Running Stable Diffusion fully in browser with WebGPU (mlc.ai) |
|
42 points by crowwork on March 8, 2023 | past | 19 comments
|
| | ML Compilation (mlc.ai) |
|
2 points by sonabinu on July 3, 2022 | past
|
|
|
More |