I'm a software engineer by training (NUST, 2025) and an AI engineer by trade. My work sits at the
intersection of shipped systems and
published research — voice AI pipelines, agentic LLMs, and
reinforcement-learning policies for wireless networks.
I've spent time benchmarking PPO in cart-pole and pendulum environments, integrating human advisory
signals into RL training loops at McMaster, and training multi-agent DRL frameworks for CR-NOMA IoT —
work that turned into two IEEE publications.
Day-to-day, I build: voice agents on VAPI + Deepgram, RAG retrieval layers with pgvector, and LLM
orchestration services in FastAPI.
Selected work
Projects, shipped and studied.
§ 02
01
RAG-Based Banking Assistant
Domain-specific LLM assistant for banking, built on LLaMA-3.2-3B with LoRA fine-tuning, FAISS retrieval, and real-time document ingestion.
Privacy-preserving ranking system trained with federated averaging on the ANTIQUE QA dataset, matching centralized baselines without pooling user data.
Multiagent Reinforcement Learning for Joint Spectrum and Energy Optimization in CR-NOMA Enabled Internet of Unmanned Agents
Saleha Ahmed, Muhammad Uzair, Syed Asad Ullah, et al.
IEEE Internet of Things Journal, 2025
A cooperative multi-agent DRL framework for CR-NOMA IoT, where distributed agents jointly learn spectrum access and power-control policies under partial observability.
Energy Efficient Uplink Communications for Wireless Powered Networks with EH Diversity: A DRL-driven Strategy
Saleha Ahmed, Muhammad Uzair, Syed Asad Ullah, et al.
IEEE International Conference on Communications (ICC), 2025
DRL-driven transmit-power control for energy-harvesting uplink nodes, evaluated against MRC, SC, and EGC diversity-combining schemes under Rayleigh fading.