February 2024 - Present

TinyLLM: Enabling Efficient LLM Deployment
on Resource-Constrained Devices

Tiny LLM

Working to develop techniques to enable the deployment of Large Language Models (LLMs) on low-power and resource-constrained devices. The primary focus is on exploring and implementing Quantization methods to reduce the computational and memory requirements of LLMs while maintaining their performance and accuracy.

May 2022 - June 2022

Terms of Service
Classification

Terms of Service Classification

Identify unfair Terms of Service clauses using a two-stage knowledge distillation DL algorithm on devices with minimal resources using state-of-the-art architectures like BERT.