My name is Arup Sarker, and I am working as a research assistant at the UVA biocomplexity Institute. My research focuses on high-performance data engineering, distributed computing, and deep learning—areas that are critical to the scale at which any system operates its personalization and content delivery systems.
Since joining my research group, I have specialized in optimizing the intersection of data processing and model execution. My work on the Deep Cylon project—building upon the Apache Arrow framework—focuses on maximizing AI algorithm performance through vectorization and efficient GPU utilization. This expertise in high-performance computing (HPC) is directly applicable to system's need for scalable, low-latency machine learning pipelines that power everything from recommendation engines to content encoding.
My portfolio includes several high-impact projects that demonstrate my ability to bridge the gap between theoretical research and scalable engineering:
Agentic DRC: Developed a unified distributed runtime for LLM and RAG pipelines to eliminate serialization bottlenecks and ensure reproducible, high-throughput execution.
Deep RC & Radical-Cylon: Engineered heterogeneous runtime systems that integrate data engineering with PyTorch and TensorFlow, reducing execution latency by up to 75.9 seconds and scaling to datasets of 3.5 billion rows.
Industry Foundation: My research is backed by a patent in image processing from my tenure at Samsung Research and a proven track record of mentorship, currently supervising eight students on the Deep Cylon project.
Before UVA, I worked as a Staff Engineer at Samsung R&D Institute Bangladesh Ltd. in September 2011. It has been a long journey with Samsung, and I have had multiple opportunities to work with different cutting-edge technologies that lead the market. Contributions to Android platform development, especially on Audio/Video, Frameworks, Network protocols to transmit data in mission-critical services and Gear360, and so many others. From the beginning at Samsung, all of my work is aligned with designing and developing embedded and distributed solutions that need to transmit data over networks. Even when I was a developer of a feature phone in 2011, I took care of system modules for transmitting calls and messages.
In addition, I was also responsible for developing Mission Critical Services for first responders and different security agencies. This project will replace the legacy systems of security agencies and ensure public safety in better ways by handling different emergencies. This large-scale distributed system will ensure all data communication, including Audio and Video, with the lowest possible latency. Every single second is very important for the first responder. So ensuring all KPIs for data quality is a major challenging job. We have completed the first milestone and delivered the product.
Experience: 16+ Years in the Software research and development industry.
• Programming Languages: C, C++, Python, Java, SQL, Objective C, Swift, Rust
• Machine Learning/Deep Learning: Time Series Models, Computer Vision(2D/3D vision model), NLP(Attention-based model), Transformers, LLMs, GenAI, HPC for Distributed LLMs, Agentic LLMs, PyTorch, Tensorflow
• Big Data/HPC: Apache Arrow, Cylon, Open MPI, UCX, NCCL, CUDA Runtime, Dask, cuDF, SLURM, Ray, DDP, FSDP, Vector Database(ChromaDB, FAISS)
• Agentic AI: RAG, Langchain, Langraph, Llamaindex
• Network Protocols: TLS over TCP/UDP, Session Initiation Protocol (SIP), Real-Time Transmission Protocol (RTP), Secure Real Time Transmission Protocol (SRTP), Media Burst Control Protocol (MBCP), Session Description Protocol (SDP), Real Time Streaming Protocol (RTSP), Simple Service Discovery Protocol(SSDP), Universal Peer to Peer Protocol (UPnP), etc. Model Context Protocol(MCP)
• Mobile Platform/Embedded System: iOS, Android, Tizen, MMP, SUP, Linux
• Other Technologies: Video/image Processing, Codec(HEVC, H.264, MJPEG), FFMPEG, Image Processing(OpenCV), WiFi, BT, BLE, Middleware frameworks, Web Service, Python-C++ binding with Cython
• Version Control: Git, Perforce, SVN
• Code Review: Gerrit for git, RBTools for perforce
• Project Management Tool: PLM, MPS, Jira
• Software Process Model: Agile (Scrum and Kanban), Waterfall
• Design Pattern : MVC, MVVM, VIPER
Department of Computer Science, University of Virginia, VA, USA
MSc in Computer Science;
PhD in Computer Science;
Department of Computer Science & Engineering, University of Dhaka, Bangladesh
MSc in Computer Science and Engineering;
BSc in Computer Science and Engineering;
https://www.linkedin.com/in/arup-sarker-8190212b/
https://scholar.google.com/citations?user=tWBCx3kAAAAJ&hl=en
Gear 360 camera can capture 360 image/video. It is small, sophisticated device, managed by iPhone and Android with Camera, Gallery, Integrated player, Share, Broadcasting, 3D Touch etc.
It is an Audio and Video Communication app where AMR Audio and H265 encoded video will be sent through SRTP to another user by following 3gpp spec of Mission Critical Services, mainly help first responder to handle public safety.
For Android platform and deployed to different Android smart phone
A Manager app combined with service to control Samsung Gear Smart watch
Video Player, Music Player, Frameworks, System, message for E1282T, E2350B, C3312R, S5222R ,E2222,E2220,S3770,C3010S
A mobile and service app for different NGOs to support Maternal health