NVIDIA Announces CUDA Toolkit v4.0

NVIDIA Announces CUDA Toolkit v4.0

New CUDA 4.0 Release Makes Parallel Programming Easier

Unified Virtual Addressing, GPU-to-GPU Communication and Enhanced C++ Template Libraries Enable More Developers to Take Advantage of GPU Computing

SANTA CLARA, Calif -- Feb. 28, 2011 -- NVIDIA today announced the latest version of the NVIDIA CUDA Toolkit for developing parallel applications using NVIDIA GPUs.

The NVIDIA CUDA 4.0 Toolkit was designed to make parallel programming easier, and enable more developers to port their applications to GPUs. This has resulted in three main features:

  • NVIDIA GPUDirect 2.0 Technology – Offers support for peer-to-peer communication among GPUs within a single server or workstation. This enables easier and faster multi-GPU programming and application performance.
     
  • Unified Virtual Addressing (UVA) – Provides a single merged-memory address space for the main system memory and the GPU memories, enabling quicker and easier parallel programming.
     
  • Thrust C++ Template Performance Primitives Libraries – Provides a collection of powerful open source C++ parallel algorithms and data structures that ease programming for C++ developers. With Thrust, routines such as parallel sorting are 5X to 100X faster than with Standard Template Library (STL) and Threading Building Blocks (TBB).

 

“Unified virtual addressing and faster GPU-to-GPU communication makes it easier for developers to take advantage of the parallel computing capability of GPUs,” said John Stone, senior research programmer, University of Illinois, Urbana-Champaign.

“Having access to GPU computing through the standard template interface greatly increases productivity for a wide range of tasks, from simple cashflow generation to complex computations with Libor market models, variable annuities or CVA adjustments,” said Peter Decrem, director of Rates Products at Quantifi. “The Thrust C++ library has lowered the barrier of entry significantly by taking care of low-level functionality like memory access and allocation, allowing the financial engineer to focus on algorithm development in a GPU-enhanced environment.”

 

The CUDA 4.0 architecture release includes a number of other key features and capabilities, including:

  • MPI Integration with CUDA Applications – Modified MPI implementations like OpenMPI automatically move data from and to the GPU memory over Infiniband when an application does an MPI send or receive call.
     
  • Multi-thread Sharing of GPUs – Multiple CPU host threads can share contexts on a single GPU, making it easier to share a single GPU by multi-threaded applications.
     
  • Multi-GPU Sharing by Single CPU Thread – A single CPU host thread can access all GPUs in a system. Developers can easily coordinate work across multiple GPUs for tasks such as “halo” exchange in applications.
     
  • New NPP Image and Computer Vision Library – A rich set of image transformation operations that enable rapid development of imaging and computer vision applications.
     
  • New and Improved Capabilities
    • Auto performance analysis in the Visual Profiler
    • New features in cuda-gdb and added support for MacOS
    • Added support for C++ features like new/delete and virtual functions
    • New GPU binary disassembler

 

A release candidate of CUDA Toolkit 4.0 will be available free of charge beginning March 4, 2011, by enrolling in the CUDA Registered Developer Program at www.nvidia.com/paralleldeveloper. The CUDA Registered Developer Program provides a wealth of tools, resources, and information for parallel application developers to maximize the potential of CUDA.

For more information on the features and capabilities of the CUDA Toolkit and on GPGPU applications, please visit: www.nvidia.com/cuda.

0
+ -

I hope, this year, we can get Dual GPU support in Vegas and Premiere Pro(Yeah, it wont happen, at least not this year) .

But when some company does achieve it , it will be super helpful and great !!

0
+ -

Well, i have a Sapphire HD 5450 (ati/amd) and i can tell you.... i kept on getting opengl erros and shut downs every time i started rendering in Adobe After effects..... but thats probably because my CPU isnt too powerful enough.... to put things in perspective, i was working on a 1920x1080 project, with close to ~500 layers... its supposed to be a kinetic typography.... Since i dont game, and video edit once in a blue moon, i'm really hoping what your saying comes true, i would mind investing in a used dual-gpu setup... though it'll be overkill for my use haha

0
+ -

Yipee rofl. Hey coolice as far as it goes dual GPU my be helpful, and not cost to much really. You do not have to go top of the line. You just have to get better than what you have now times two. With prices on GPU's now unless it is top of the line I bet you could get 2 for the price of 1 of the top of the line cards.

Login or Register to Comment
Post a Comment
Username:   Password: