The NVIDIA Full Stack
From CUDA Kernels to Cloud-Native AI Deployment
Failed to add items
Sorry, we are unable to add the item because your shopping cart is already at capacity.
Add to Cart failed.
Please try again later
Add to Wish List failed.
Please try again later
Remove from wishlist failed.
Please try again later
Adding to library failed
Please try again
Follow podcast failed
Please try again
Unfollow podcast failed
Please try again
Audible Standard 30-day free trial
Select 1 audiobook a month from our entire collection of titles.
Yours as long as you’re a member.
Get unlimited access to bingeable podcasts.
Standard auto renews for $8.99 a month after 30 days. Cancel anytime.
Buy for $6.40
-
Narrated by:
-
Virtual Voice
-
By:
-
Ajit Singh
This title uses virtual voice narration
Virtual voice is computer-generated narration for audiobooks.
Philosophy
The core philosophy of this book is to present the NVIDIA platform as a cohesive, end-to-end "full stack." Traditional resources often treat CUDA programming, AI frameworks, and model deployment as separate disciplines. This book dismantles those silos. I believe that a modern AI engineer must understand the entire lifecycle of an application: from the low-level CUDA kernels that execute on the hardware, to the high-level Python frameworks used for training, to the cloud-native tools required for scalable, production-grade deployment. This approach provides a holistic understanding that is essential for building efficient, robust, and maintainable systems in the real world.
Key Features
1. Full-Stack Coverage: The only book you need to understand the journey from CUDA C++ and Python kernels to production deployment with Triton and Docker.
2. Beginner to Advanced: Carefully structured to cater to both undergraduate students (B.Tech) and postgraduate specialists (M.Tech), as well as professional developers.
3. Globally Relevant: The technologies covered (CUDA, PyTorch, Docker, MLOps) are industry standards, making the curriculum compatible with international university syllabi.
4. Code-Intensive: Rich with working code examples, practical exercises, and a complete, explained capstone project.
5. Focus on Optimization: Dedicated chapters on profiling with Nsight tools and inference optimization with TensorRT, teaching the critical skill of making AI applications fast and efficient.
To Whom This Book Is For
1. B.Tech/M.Tech Computer Science Students: An ideal textbook or supplementary resource for courses on Parallel Computing, High-Performance Computing, AI/ML, and Cloud Computing.
2. Aspiring AI/ML Engineers: Provides the essential hands-on skills required for a career in building and deploying AI systems.
3. Data Scientists: For those who want to move beyond notebooks and learn how to accelerate their data pipelines and deploy models at scale using RAPIDS and Triton.
4. Software Developers & Researchers: A practical guide for professionals looking to leverage GPU acceleration for their applications, whether in scientific computing, finance, or any other domain.
Disclaimer: Earnest request from the Author.
Kindly go through the table of contents and refer kindle edition for a glance on the related contents.
Thank you for your kind consideration!
People who viewed this also viewed...
No reviews yet