speaker detail

Anshuman Mishra

Machine Learning Engineer

company logo

Anshuman Mishra is a talented Machine Learning Engineer and Google Developer Expert (GDE) in ML (GenAI) with a diverse background in software engineering and data science. Currently, he is contributing his expertise as a full-time Machine Learning Engineer at Flip AI, where he plays a pivotal role in developing cutting-edge models and software services. Despite being the youngest member of the team, Anshuman is valued for his contributions and enjoys equal respect and responsibilities.

This session offers an in-depth exploration of leveraging CUDA to optimize NVIDIA hardware, essential for the explosive growth of Generative AI applications. Generative AI models, which include image and text generation and code completion, rely heavily on accelerated hardware for efficient training and inference. While high-level frameworks like PyTorch and TensorFlow simplify the process, true optimization and control are unlocked through CUDA, NVIDIA’s low-level compiler interfaces directly with GPUs.

The session begins with thoroughly reviewing C programming fundamentals, ensuring a solid base. It then demystifies core CUDA concepts, including threads, blocks, grids, and memory hierarchies, teaching participants to think in parallel for efficient GPU utilization. The course is designed to be highly interactive, with practical sessions guiding learners through writing their kernels, the essential workhorses of CUDA programs. This hands-on approach provides participants with real-world experience in parallel programming, ensuring they understand and effectively utilize the power of parallel processing in Generative AI.

Read More

Managing and scaling ML workloads have never been a bigger challenge in the past. Data scientists are looking for collaboration, building, training, and re-iterating thousands of AI experiments. On the flip side ML engineers are looking for distributed training, artifact management, and automated deployment for high performance

Read More

Managing and scaling ML workloads have never been a bigger challenge in the past. Data scientists are looking for collaboration, building, training, and re-iterating thousands of AI experiments. On the flip side ML engineers are looking for distributed training, artifact management, and automated deployment for high performance

Read More

We use cookies essential for this site to function well. Please click to help us improve its usefulness with additional cookies. Learn about our use of cookies in our Privacy Policy & Cookies Policy.

Show details