About

Jongseok Park


Hello! I'm Jongseok Park, a second-year Master's student in Electrical and Computer Engineering (ECE) at Seoul National University, advised by Prof. Kyunghan Lee. I'm part of the Networked Computing and ML lab, which aims to innovate mobile and networked computing with hyper-connectivity, expanding the capabilities of power-limited mobile platforms. My interest lies in designing computation systems, scheduling algorithms, and acceleration hardware that enable the efficient use of new, high-cost applications, such as machine learning, by leveraging the characteristics of the target environment.

Check out the Gallery and Personal Projects for my past activities!

Publications

ASPEN: Breaking Operator Barriers for Efficient Parallel Execution of Deep Neural Networks (NeurIPS'23)

Overview

ASPEN is a novel parallel computation system for DNNs that allows fine-grained dynamic execution of DNNs. ASPEN removes synchronization barriers of tensor operators and expresses DNNs in dataflow graphs of fine-grained tiles, exposing novel computation opportunities across operators. ASPEN exploits these opportunities in runtime by dynamically locating and scheduling them in a distributed and asynchronous manner.

ASPEN enables opportunistic parallelism, a new class of parallelism for DNNs that is unavailable in operator-based approaches. ASPEN’s graph-wide parallel scheduling scope paired with dynamic, distributed runtime enables highly efficient resource utilization, scalability, and load balancing. Also, ASPEN achieves higher data reuse and reduced communications by letting each resource asynchronously traverse depthwise in the DNN graph.

Links & Materials

mGEMM: Low-latency Convolution with Minimal Memory Overhead Optimized for Mobile Devices (MobiSys'22)

Overview

mGEMM is a convolution computation algorithm that focuses on the low-latency requirements of mobile AI applications. mGEMM removes the memory duplication and reordering overhead (im2col) of existing matrix multiplication (GEMM) based convolution algorithms. mGEMM achieves this by adding filter spatial dimensions of convolutions in the assembly-level GEMM computation kernel, allowing the reuse of ARM NEON vector registers over GEMM iterations on different filter spatial dimensions.

mGEMM achieves improved latency, memory usage, and energy consumption in real-world mobile devices, and can be used as a drop-in replacement for existing convolution kernels. 

Links & Materials

Jongseok ParkDepartment of Electrical & Computer EngineeringSeoul National University1151-2, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of KoreaLast updated November 26, 2023