Credits & Acknowledgments

LiteEdgeAI is made possible by two essential open-source projects. Without them, browser-based AI benchmarking would not exist.

MLC-AI / WebLLM ↗

#1 Core Engine

What it does:

WebLLM is the entire foundation of this project. It enables running large language models completely in your browser using WebGPU for GPU acceleration. Without WebLLM, none of the benchmarking, model loading, or inference would be possible.

How we use it:

  • Loading quantized LLM models directly in the browser
  • Running GPU-accelerated inference using WebGPU
  • Measuring tokens per second, load times, and performance metrics
  • Detecting GPU capabilities and system specifications
License: Apache 2.0Version: 0.2.80Creators: MLC-AI Team

Hugging Face ↗

#2 Model Platform

What it does:

Hugging Face is the platform hosting and distributing all the AI models used in our benchmarks. They provide the infrastructure to access pre-quantized models optimized for browser execution. Without their model repository and CDN, we couldn't load models efficiently.

How we use it:

  • Downloading MLC-quantized model files on-demand
  • Accessing models from Meta, Microsoft, Alibaba, and others
  • Leveraging their global CDN for fast model delivery

Models we benchmark:

TinyLlama 1.1B
TinyLlama team
Llama 3.2 1B
Meta AI
Phi-3 Mini
Microsoft
Qwen 2.5 1.5B
Alibaba Cloud
Platform: Model hosting & distributionFounded: 2016

These two projects are the backbone of LiteEdgeAI. We are deeply grateful to the MLC-AI team and Hugging Face for building and maintaining the infrastructure that makes browser-based AI benchmarking possible. All open-source licenses are respected and followed.