EdgeCompress is a graduation project developed by senior Computer Science students from Capital University, Egypt.
Our work focuses on compressing Large Language Models (LLMs) to make them efficient enough to run on edge devices with limited computational resources.
Large Language Models typically require significant memory, storage, and computational power. This makes them difficult to deploy on edge hardware such as embedded systems, IoT devices, and low-power GPUs.
Our project explores different model compression techniques to reduce the size and resource requirements of LLMs while maintaining acceptable performance.
We investigate multiple compression approaches, including:
After compression, the models are evaluated on edge computing environments to determine:
This organization hosts:
Our goal is to enable efficient deployment of LLMs on edge devices, making advanced AI models more accessible in real-world and resource-constrained environments.