Found 12 packages in 0.03 seconds
Kernel Operations on GPU or CPU, with Autodiff, without Memory Overflows
The 'KeOps' library lets you compute generic reductions of very large arrays whose entries are given by a mathematical formula with CPU and GPU computing support. It combines a tiled reduction scheme with an automatic differentiation engine. It is perfectly suited to the efficient computation of Kernel dot products and the associated gradients, even when the full kernel matrix does not fit into the GPU memory.
'GPU'/CPU Benchmarking in Debian-Based Systems
'GPU'/CPU Benchmarking on Debian-package based systems This package benchmarks performance of a few standard linear algebra operations (such as a matrix product and QR, SVD and LU decompositions) across a number of different 'BLAS' libraries as well as a 'GPU' implementation. To do so, it takes advantage of the ability to 'plug and play' different 'BLAS' implementations easily on a Debian and/or Ubuntu system. The current version supports - 'Reference BLAS' ('refblas') which are un-accelerated as a baseline - Atlas which are tuned but typically configure single-threaded - Atlas39 which are tuned and configured for multi-threaded mode - 'Goto Blas' which are accelerated and multi-threaded - 'Intel MKL' which is a commercial accelerated and multithreaded version. As for 'GPU' computing, we use the CRAN package - 'gputools' For 'Goto Blas', the 'gotoblas2-helper' script from the ISM in Tokyo can be used. For 'Intel MKL' we use the Revolution R packages from Ubuntu 9.10.