Skip to content
#

cuda

Here are 2,561 public repositories matching this topic...

numba
gmarkall
gmarkall commented Nov 3, 2020

PR #6447 adds a public API to get the maximum number of registers per thread (numba.cuda.Dispatcher.get_regs_per_thread()). There are other attributes that might be nice to provide - shared memory per block, local memory per thread, const memory usage, maximum block size.

These are all available in the FuncAttr named tuple: https://github.com/numba/numba/blob/master/numba/cuda/cudadrv/drive

revans2
revans2 commented Nov 23, 2020

Spark is really inconsistent in how it handles some values like -0.0 vs 0.0 and the various NaN values that are possible. I don't expect cuDF to be aware of any of this, but I would like the ability to work around it in some cases by treating the floating point value as if it were just a bunch of bits. To me logical_cast feels like the right place to do this, but floating point values are

miguelusque
miguelusque commented Dec 6, 2020

Is your feature request related to a problem? Please describe.
While porting some code from SKL to cuML, I have noticed the following:

SKL:
from sklearn.model_selection import train_test_split
cuML:
from cuml.preprocessing.model_selection import train_test_split

If I try to do from cuml.model_selection import train_test_split, the following error is displayed:
`ModuleNotFoundE

futhark

Improve this page

Add a description, image, and links to the cuda topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the cuda topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.