-
Archives
- June 2020
- May 2020
- March 2020
- August 2019
- June 2019
- April 2019
- March 2019
- May 2018
- April 2018
- October 2017
- September 2017
- July 2017
- April 2017
- February 2017
- January 2017
- December 2016
- August 2016
- July 2016
- June 2016
- April 2016
- March 2016
- February 2016
- January 2016
- October 2015
- September 2015
- July 2015
- June 2015
- May 2015
- April 2015
- November 2014
- October 2014
- September 2014
- June 2014
- May 2014
- April 2014
- March 2014
- February 2014
- January 2014
- December 2013
- November 2013
- October 2013
- September 2013
- August 2013
- July 2013
- June 2013
- April 2013
- March 2013
- February 2013
- January 2013
- September 2012
- May 2012
- April 2012
- January 2012
- December 2011
- September 2011
- August 2011
- July 2011
-
Meta
Tag Archives: gpu
NVIDIA Tesla K20 GPU Accelerator (Kepler GK110) Up Close
NVIDIA’s Tesla K20 GPU is currently the de facto standard for high-performance heterogeneous computing. Based upon the Kepler GK110 architecture, these are the GPUs you want if you’ll be taking advantage of the latest advancements available in CUDA 5.0 and … Continue reading
GPU Memory Types – Performance Comparison
This post is Topic #3 (part 1) in our series Parallel Code: Maximizing your Performance Potential. CUDA devices have several different memory spaces: Global, local, texture, constant, shared and register memory. Each type of memory on the device has its … Continue reading
Optimize CUDA Host/Device Transfers
This post is Topic #2 (part 2) in our series Parallel Code: Maximizing your Performance Potential. In my previous post, CUDA Host/Device Transfers and Data Movement, I provided an introduction into the bottlenecks associated with host/device transfers and data movement. … Continue reading
CUDA Host/Device Transfers and Data Movement
This post is Topic #2 (part 1) in our series Parallel Code: Maximizing your Performance Potential. In post #1, I discussed a few ways to optimize the performance of your application via controlling your threads and provided some insight as … Continue reading
CUDA Parallel Thread Management
This post is Topic #1 in our series Parallel Code: Maximizing your Performance Potential. Regardless of the environment or architecture you are using, one thing is certain: you must properly manage the threads running in your application to optimize performance. This … Continue reading
Parallel Code: Maximizing your Performance Potential
No matter what the purpose of your application is, one thing is certain. You want to get the most bang for your buck. You see research papers being published and presented making claims of tremendous speed increases by running algorithms … Continue reading
5 Easy First Steps on GPUs – Accelerating Your Applications
This week NVIDIA provided a tutorial outlining first steps for GPU acceleration using OpenACC and CUDA. This was offered as part of the “GPUs Accelerating Research” week at Northeastern University and Boston University. After attending, it seemed appropriate to review … Continue reading
NVIDIA Tesla K10 GPU Accelerator (Kepler GK104) Up Close
NVIDIA is now shipping their 4.58 TFLOPS single-precision floating point GPUs. The Tesla K10 GPU Accelerators, based upon the Kepler GK104 architecture, are the first Teslas available from this new generation of products. They are designed for single-precision float-point applications, … Continue reading
Inside NVIDIA Kepler – Live from GTC 2012
Compute performance has been exponentially increasing for the entirety of your life – it doesn’t matter what your age is. This week at NVIDIA’s GTC 2012 conference, we’ve seen that GPUs are still leading the charge. The new NVIDIA “Kepler” … Continue reading
nvidia-smi: Control Your GPUs
This post was last updated on 2018-11-05 Most users know how to check the status of their CPUs, see how much system memory is free, or find out how much disk space is free. In contrast, keeping tabs on the … Continue reading