CV

Basics

Name Srinivasan Subramaniyan
Label Ph.D. Candidate in Computer Engineering
Email subramaniyan.4@osu.edu
Phone +1 740 274 2814
Url https://srinivasans74.github.io/
Summary Ph.D. candidate in Electrical and Computer Engineering at The Ohio State University, focusing on GPU scheduling, data center efficiency, and LLM/AI systems. Awarded Best Paper honors at EMSOFT 2025 and VLSID 2022. Seeking internships in computer engineering, software systems, or high-performance computing.

Work

  • 2022.05 - 2022.08
    Research Intern
    Advanced Micro Devices (AMD)
    Optimized the scheduling of GP-GPU kernels to accelerate graph-based applications, improving performance and efficiency.
    • Worked on GPU workload scheduling optimizations for graph-based applications.
    • Enhanced compute efficiency and resource utilization for AMD research workloads.
  • 2019.01 - 2021.08
    Junior Research Fellow
    Centre for Heterogeneous and Intelligent Processing Systems (CHIPS)
    Conducted research on FPGA-based acceleration for communication and ML workloads.
    • Performed design-space exploration for NB-LDPC codes on FPGAs (SIPS ’20, IEEE Design & Test ’22).
    • Developed hardware accelerators for sparse matrix multiplication (VLSID ’22).

Education

  • 2021.08 - Present

    Columbus, Ohio, USA

    M.S. + Ph.D.
    The Ohio State University
    Computer Engineering
    • Computer Architecture
    • Embedded Systems
    • Parallel and Distributed Systems
    • Operating Systems
    • High-Performance Computing
    • Reinforcement Learning and Machine Learning
    • FPGA/SoC Design and Performance Modeling

Skills

Programming Languages
C/C++
Python
Bash
x86/ARM/RISC-V Assembly
Hardware Design & Verification
Verilog
SystemVerilog
FPGA/SoC Design
Hardware Simulation & Debugging
Parallel & Distributed Computing
OpenCL
CUDA
OpenMP
HIP
MPI
Optimization & Modeling
Gurobi
PuLP
Simulink
Performance Profilers (gprof, perf, NVProf, Nsight)
Development & Tools
Git
Linux Kernel Modules
Docker
Kubernetes
vLLM

Awards

  • 2025.10.01
    Outstanding Paper Award, EMSOFT 2025
    ACM/IEEE Embedded Systems Week
    Recognized for the paper 'FC-GPU: Feedback Control GPU Scheduling for Real-time Embedded Systems'.
  • 2022.01.10
    Best Paper Award, VLSID 2022
    International Conference on VLSI Design and Embedded Systems
    Received the A.K. Choudhary Best Paper Award for FPGA accelerator design for sparse matrix multiplication.
  • 2025.09.15
    EMSOFT Travel Grant Award
    Embedded Systems Week
    Awarded travel support for presenting research at EMSOFT 2025.
  • 2025.05.01
    BurnLin Travel Grant Award
    The Ohio State University
    Awarded three consecutive years (2023–2025) for research excellence in embedded and parallel systems.
  • 2018.05.01
    Amrita Scholarship
    Amrita Vishwa Vidyapeetham
    Merit-based undergraduate scholarship recognizing academic excellence.

Publications

Projects

  • FC-GPU: Feedback-Control GPU Scheduling
    Developed FC-GPU, the first feedback-control GPU scheduling framework for real-time systems using a MIMO controller to dynamically adapt task rates, reducing deadline misses by 2%.
    • Best Paper Candidate, EMSOFT 2025
    • Reduced deadline misses by 2% on RTX 3090 and MI100.
  • CapLLM: Power-Capping for LLM Data Centers
    Designed CapLLM, a power-capping framework for LLM-serving data centers that minimizes performance violations while improving energy efficiency.
    • Dynamic GPU power management
    • Improved SLA compliance
  • CorrGPU: Correlation-Aware GPU Scheduling
    Proposed CorrGPU, a scheduling algorithm that consolidates correlated workloads to reduce contention and lower CapEx by 20.88%.
    • Correlation-aware scheduling
    • Reduced CapEx by 20.88%
  • CapGPU: Coordinated CPU–GPU Power Capping
    Implemented CapGPU, a coordinated CPU–GPU power-capping strategy improving inference throughput by 8–20% while maintaining latency SLOs.
    • 8–20% throughput improvement
    • Maintained SLO compliance
  • GPUColo: GPU Co-Location Framework
    Built GPUColo, a co-location framework that enables training and inference workloads to share GPUs, saving up to 74.9% of GPUs with strict SLO compliance.
    • 74.9% GPU savings
    • Reduced CapEx with SLO guarantees

Languages

English
Fluent
Tamil
Native
Hindi
Professional working proficiency