• Improved runtime of C++ Vulkan GPU based DO-178 compliant AI inference engine from 150ms to 70ms,
by improving host program, optimizing host device memory transfers and improving the GPU kernel
• Wrote Vulkan HLSL GPU Compute kernel to perform mean pooling operation in neural network, which
allowed the AI inference engine to successfully run the vgg16 and vgg18 neural networks
• Made presentations and gave talks internally in the company about Vulkan and GPU compute, focusing on
how to design Vulkan host applications and high performance high parallelism GPU kernels
• Created testing tools to measure kernel performance and correctness automatically for new iterations