Sign in with
Sign up | Sign in

Nvidia, Udacity Offer Free Parallel Programming Courses

By - Source: Nvidia | B 10 comments

Nvidia is expanding its academic GPU programming classes. This time, however, it is offering a free class.

The company said it is partnering with Udacity to provide insight in parallel programming with a class that is designed to hold material worth about one week of studying. This specific class is described as an introduction to the development model and covers CUDA coding with "a series of image processing algorithms, such as you might find in Photoshop or Instagram."

Students can test their code on the company's recently announced cloud GPUs. The class is taught by John Owens, associate professor of electrical and computer engineering at the University of California, Davis, as well as David Luebke senior director of graphics research at Nvidia.

Anyone can sign up for the course, so check it out if you're interested.

Nvidia Udacity Courses

Contact Us for News Tips, Corrections and Feedback           

Display 10 Comments.
This thread is closed for comments
  • -7 Hide
    warezme , October 23, 2012 9:42 PM
    what?, my cock thinks of gpu pron
  • 7 Hide
    Estix , October 23, 2012 11:10 PM
    It seems like Nvidia is realizing what Microsoft did a long time ago; you want people to code for your platform, even if it means spending a bit to attract them.
  • 2 Hide
    Old_Fogie_Late_Bloomer , October 24, 2012 12:21 AM
    Coursera is offering a six-week program on this subject that's supposed to start this month...I'm already signed up and I can't wait for it to start. :)
  • 4 Hide
    A Bad Day , October 24, 2012 12:33 AM
    Your standard software company's disconnected management:

    "Hm, do we either send some of the programmers away for at least week while still paying them, or do we launch the software by Thanksgiving?

    You know what, F*** parallel computing or whatever it's called, I need something on the deadline!"
  • 0 Hide
    idono , October 24, 2012 1:02 AM
    Signedup right now.
  • 2 Hide
    morstern , October 24, 2012 1:32 AM
    "Hm, do we either send some of the programmers away for at least week while still paying them, or do we launch the software by Thanksgiving?..."

    I think yes we invest in current and future training for our programmers so that we stay relavent.
  • -2 Hide
    falchard , October 24, 2012 5:39 AM
    Problem is its based on CUDA and not OpenCL or DirectCompute. Who cares about CUDA?
  • 1 Hide
    devBunny , October 24, 2012 4:00 PM
    falchardProblem is its based on CUDA and not OpenCL or DirectCompute. Who cares about CUDA?

    To a newbie, any experience in parallel programming is invaluable. If the course has enough theory in it then much of the acquired skill will transfer to OpenCL and DirectCompute. Similarly, the course is based on image processing so it might be tempting to say "Who cares?" if you want to learn to use GPUGP for, say, simulations, but, again, valuable lessons can be learned.
  • 0 Hide
    Anonymous , October 25, 2012 9:08 PM
    CUDA is easier to learn than OpenCL, and it's even recommended to learn CUDA first to get familiar with parallel programming concepts, even if you later want to move on to OpenCL, etc
  • 0 Hide
    jowens , October 29, 2012 8:21 PM
    As you might imagine, we had a lot of conversations about the right way to teach parallel computing using GPUs. Three things to note, my opinions here only. I think anyone who's programmed in both would agree that CUDA is simpler, particularly at the outset; setting up a "hello, world" kind of program on CUDA is just a lot easier than with OpenCL. Also, I think once you understand either CUDA or OpenCL, it's a piece of cake to pick up the other one; learning the first one is a lot more challenging than learning the second one, so I imagine anyone that takes the course and then wants to pick up OpenCL will have little difficulty. Finally, the ecosystem around CUDA is just a lot more developed than around OpenCL, so if students want/need to go grab a parallel sort routine or scan or other primitives, they'll have a lot easier time in CUDA than in OpenCL, which is important as they start writing more complex programs.

    In my research group we program in both CUDA and OpenCL, and we get significant funding from both Intel and AMD who are both OpenCL-focused; but I think I would accurate in saying that strictly from a technical perspective, all my grad students would prefer to code in CUDA if they had to choose.