CV
Basics
| Name | Srinivasan Subramaniyan |
| Label | Ph.D. Candidate in Computer Engineering |
| subramaniyan.4@osu.edu | |
| Phone | +1 740 274 2814 |
| Url | https://srinivasans74.github.io/ |
| Summary | Ph.D. candidate in Electrical and Computer Engineering at The Ohio State University, focusing on GPU scheduling, data center efficiency, and LLM/AI systems. Awarded Best Paper honors at EMSOFT 2025 and VLSID 2022. Seeking internships in computer engineering, software systems, or high-performance computing. |
Work
-
2022.05 - 2022.08 Research Intern
Advanced Micro Devices (AMD)
Optimized the scheduling of GP-GPU kernels to accelerate graph-based applications, improving performance and efficiency.
- Worked on GPU workload scheduling optimizations for graph-based applications.
- Enhanced compute efficiency and resource utilization for AMD research workloads.
-
2019.01 - 2021.08 Junior Research Fellow
Centre for Heterogeneous and Intelligent Processing Systems (CHIPS)
Conducted research on FPGA-based acceleration for communication and ML workloads.
- Performed design-space exploration for NB-LDPC codes on FPGAs (SIPS ’20, IEEE Design & Test ’22).
- Developed hardware accelerators for sparse matrix multiplication (VLSID ’22).
Education
-
2021.08 - Present Columbus, Ohio, USA
M.S. + Ph.D.
The Ohio State University
Computer Engineering
- Computer Architecture
- Embedded Systems
- Parallel and Distributed Systems
- Operating Systems
- High-Performance Computing
- Reinforcement Learning and Machine Learning
- FPGA/SoC Design and Performance Modeling
Skills
| Programming Languages | |
| C/C++ | |
| Python | |
| Bash | |
| x86/ARM/RISC-V Assembly |
| Hardware Design & Verification | |
| Verilog | |
| SystemVerilog | |
| FPGA/SoC Design | |
| Hardware Simulation & Debugging |
| Parallel & Distributed Computing | |
| OpenCL | |
| CUDA | |
| OpenMP | |
| HIP | |
| MPI |
| Optimization & Modeling | |
| Gurobi | |
| PuLP | |
| Simulink | |
| Performance Profilers (gprof, perf, NVProf, Nsight) |
| Development & Tools | |
| Git | |
| Linux Kernel Modules | |
| Docker | |
| Kubernetes | |
| vLLM |
Awards
- 2025.10.01
Outstanding Paper Award, EMSOFT 2025
ACM/IEEE Embedded Systems Week
Recognized for the paper 'FC-GPU: Feedback Control GPU Scheduling for Real-time Embedded Systems'.
- 2022.01.10
Best Paper Award, VLSID 2022
International Conference on VLSI Design and Embedded Systems
Received the A.K. Choudhary Best Paper Award for FPGA accelerator design for sparse matrix multiplication.
- 2025.09.15
EMSOFT Travel Grant Award
Embedded Systems Week
Awarded travel support for presenting research at EMSOFT 2025.
- 2025.05.01
BurnLin Travel Grant Award
The Ohio State University
Awarded three consecutive years (2023–2025) for research excellence in embedded and parallel systems.
- 2018.05.01
Amrita Scholarship
Amrita Vishwa Vidyapeetham
Merit-based undergraduate scholarship recognizing academic excellence.
Publications
-
2025.10.01 FC-GPU: Feedback Control GPU Scheduling for Real-time Embedded Systems
EMSOFT 2025
Proposed FC-GPU, a feedback-control framework for real-time GPU scheduling that reduces deadline misses by 2% using MIMO-based control.
-
2025.09.01 Power Capping of GPU Servers for Machine Learning Inference Optimization
ICPP 2025
Proposed CapGPU, a coordinated CPU–GPU power-capping framework that improves inference throughput by up to 20% under latency constraints.
-
2025.08.01 Exploiting ML Task Correlation in the Minimization of Capital Expense for GPU Data Centers
IEEE IPCCC 2025
Introduced CorrGPU, a correlation-aware scheduling algorithm that reduces CapEx by 20.88% in large-scale ML workloads.
-
2024.07.01 Latency-Guaranteed Co-Location of Inference and Training for Reducing Data Center Expenses
IEEE ICDCS 2024
Developed GPUColo, a co-location framework that enables training and inference sharing on GPUs, saving up to 74.9% of GPU resources.
Projects
-
FC-GPU: Feedback-Control GPU Scheduling
Developed FC-GPU, the first feedback-control GPU scheduling framework for real-time systems using a MIMO controller to dynamically adapt task rates, reducing deadline misses by 2%.
- Best Paper Candidate, EMSOFT 2025
- Reduced deadline misses by 2% on RTX 3090 and MI100.
-
CapLLM: Power-Capping for LLM Data Centers
Designed CapLLM, a power-capping framework for LLM-serving data centers that minimizes performance violations while improving energy efficiency.
- Dynamic GPU power management
- Improved SLA compliance
-
CorrGPU: Correlation-Aware GPU Scheduling
Proposed CorrGPU, a scheduling algorithm that consolidates correlated workloads to reduce contention and lower CapEx by 20.88%.
- Correlation-aware scheduling
- Reduced CapEx by 20.88%
-
CapGPU: Coordinated CPU–GPU Power Capping
Implemented CapGPU, a coordinated CPU–GPU power-capping strategy improving inference throughput by 8–20% while maintaining latency SLOs.
- 8–20% throughput improvement
- Maintained SLO compliance
-
GPUColo: GPU Co-Location Framework
Built GPUColo, a co-location framework that enables training and inference workloads to share GPUs, saving up to 74.9% of GPUs with strict SLO compliance.
- 74.9% GPU savings
- Reduced CapEx with SLO guarantees
Languages
| English | |
| Fluent |
| Tamil | |
| Native |
| Hindi | |
| Professional working proficiency |