-
Updated
Sep 16, 2020 - Makefile
cuda
Here are 2,548 public repositories matching this topic...
-
Updated
Dec 3, 2020 - Shell
Problem:
catboost version: 0.23.2
Operating System: all
Tutorial: https://github.com/catboost/tutorials/blob/master/custom_loss/custom_metric_tutorial.md
Impossible to use custom metric (С++).
Code example
from catboost import CatBoost
train_data = [[1, 4, 5, 6],
-
Updated
Nov 25, 2020 - Python
-
Updated
Dec 3, 2020 - C++
-
Updated
Dec 3, 2020 - C++
-
Updated
Nov 29, 2020 - Go
Is your feature request related to a problem? Please describe.
from time series analysis, need to support spearman correlation matrix calculation in cuDF
Describe the solution you'd like
similar like pandas.DataFrame().corr(method='spearman')
Additional context
is it possible to let me know this feature adding roadmap? big thanks !
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
Names map
and input
are exchanged mistakenly. By sense of Preconditions
paragraph they have to be exchanged I suppose, because there is no problem when map
and result
coincide (in current context).
-
Updated
Dec 1, 2020 - C++
The marker no_bad_cuml_array_check
was necessary in 0.16 to avoid asserting on bad uses of CumlArray
in specific tests. As of 0.17, this is no longer necessary.
This marker should be removed from pytest.ini
as well as any tests that used it. A quick search shows this is used in 2 tests in test_incrementa_pca.py
-
Updated
Sep 11, 2018 - C++
I often use -v
just to see that something is going on, but a progress bar (enabled by default) would serve the same purpose and be more concise.
We can just factor out the code from futhark bench
for this.
-
Updated
Dec 2, 2020 - Python
-
Updated
Sep 29, 2020 - Jupyter Notebook
-
Updated
Jul 22, 2020 - C++
Thank you for this fantastic work!
Could it be possible the fit_transform() method returns the KL divergence of the run?
Thx!
-
Updated
Dec 2, 2020 - Python
-
Updated
Nov 22, 2020 - Python
-
Updated
Nov 23, 2020 - C++
-
Updated
Dec 1, 2020 - Python
Improve this page
Add a description, image, and links to the cuda topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the cuda topic, visit your repo's landing page and select "manage topics."
PR #6447 adds a public API to get the maximum number of registers per thread (
numba.cuda.Dispatcher.get_regs_per_thread()
). There are other attributes that might be nice to provide - shared memory per block, local memory per thread, const memory usage, maximum block size.These are all available in the
FuncAttr
named tuple: https://github.com/numba/numba/blob/master/numba/cuda/cudadrv/drive