Dr. Dobb's is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them. Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.


Channels ▼



site Search Results

Results for: farber

Sort by: Relevance | Date

The Dr. Dobb's Developer Library DVD Order the NEW Discounted Dr. Dobb's Developer Library DVD 6
Purchase the fully searchable DVD for $59.95 - a 60% discount! Features 21 years of Dr. Dobb's Journal, 15 years of Sys Admin magazine, 14+ years of C/C++ Users Journal, 1 year worth of Dr. Dobb's Digest, podcasts, videos and more! Order Now.

A Massively Parallel Stack for Data Allocation

Dr. Dobb's Journal,
October 01, 2013

A fast, constant, type memory allocator and parallel stack are essential for initiating kernel launches from the CUDA device - Parallel

A Robust Histogram for Massive Parallelism

Dr. Dobb's Journal,
September 24, 2013

Preserving highly parallel performance when every thread is simultaneously trying to increment a single object - Parallel

CUDA: Unifying Host/Device Interactions with a Single C++ Macro

Dr. Dobb's Journal,
September 16, 2013

A general method to move data transparently between the host and the CUDA device. - Parallel

Atomic Operations and Low-Wait Algorithms in CUDA

Dr. Dobb's Journal,
August 20, 2013

Used correctly, atomic operations can help implement a wide range of generic data structures and algorithms in the massively threaded GPU programming environment. However, incorrect usage can turn massively parallel GPUs into poorly performing sequential processors. - Parallel

Exceeding Supercomputer Performance with Intel Phi

Dr. Dobb's Journal,
June 04, 2013

Using MPI on inexpensive clusters of Intel Xeon Phi coprocessors can produce results that exceed the performance of today's high-end supercomputers. - Parallel

Numerical and Computational Optimization on the Intel Phi

Dr. Dobb's Journal,
March 19, 2013

How tuning functions for large data sets and profiling the results gets most of the benefits of the Phi's 60 cores without hand wringing and late-night hacking. - Parallel

Getting to 1 Teraflop on the Intel Phi Coprocessor

Dr. Dobb's Journal,
March 12, 2013

The key to truly high performance with the Phi coprocessor is to express sufficient parallelism and vector capability to fully utilize the device. Here is a timing framework that enables you to measure and optimize performance and push it past 1 teraflop. - Parallel

Programming the Xeon Phi

Dr. Dobb's Journal,
February 02, 2013

A series of articles on getting the best performance out of the new Intel Xeon Phi coprocessor - Parallel

Comparing OpenCL, CUDA, and OpenACC [video]

Dr. Dobb's Journal,
January 01, 2013

Rob Farber takes you on a tour of the paths to massively parallel x86, MultiGPU, and CPU+GPU applications. - Parallel

CUDA vs. Phi: Phi Programming for CUDA Developers

Dr. Dobb's Journal,
December 17, 2012

Both CUDA and Phi coprocessors provide high degrees of parallelism that can deliver excellent application performance. For the most part, CUDA programmers with existing application code have already written their software so it can run well on Phi coprocessors. However, additional work may be required to achieve the highest possible performance. - Parallel

1 2 3 4 5 Next