Umut Onur Yaşar
AI/ML Engineer · Deep Learning & Computer Vision
I work at the intersection of model efficiency and real-world deployment — designing, training, and optimizing neural networks with a focus on speed-accuracy trade-offs.
My background in Electrical and Electronics Engineering gives me a low-level intuition that complements high-level ML research: from CUDA kernel optimization to transformer architecture design, I can reason across the full stack.
Current Work
KD-CIFAR10: Knowledge Distillation Ablation Study (Paper in preparation) Systematic ablation of Logit-KD vs Feature-KD for compressing ResNet-50 (23.5M) into ResNet-18 (11.2M) on CIFAR-10. Key finding: architecture gap is the primary distillation bottleneck — Logit-KD reaches 95.47% (+0.50pp over baseline) once the stem is corrected for small-input datasets. T=4 optimal across all runs.
Research Interests
- Model Compression — knowledge distillation, pruning, quantization
- Object Detection — transformer-based architectures (RT-DETR, DETR variants)
- Efficient Inference — CUDA optimization, edge deployment
- Representation Learning — feature alignment, intermediate supervision
Selected Projects
| Project | Stack | Description |
|---|---|---|
| KD-CIFAR10 | PyTorch · ResNet · CIFAR-10 | Logit-KD vs Feature-KD ablation for ResNet-50→18 compression |
| RT-DETR Knowledge Distillation | PyTorch · CUDA · COCO | Teacher-student framework for real-time object detection |
| Real-Time Object Detection | Python · YOLO · OpenCV | Real-time detection pipeline |
| Signal Processing Toolbox | C++ · Qt · FFT | Filtering, Fourier analysis, waveform visualization |
| RF Propagation & Coverage Tool | Python | RF signal simulation and visualization |
