-
Updated
Jun 23, 2021 - Makefile
CUDA

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
Here are 2,947 public repositories matching this topic...
-
Updated
Jul 19, 2021 - Shell
Problem: the approximate method can still be slow for many trees
catboost version: master
Operating System: ubuntu 18.04
CPU: i9
GPU: RTX2080
Would be good to be able to specify how many trees to use for shapley. The model.predict and prediction_type versions allow this. lgbm/xgb allow this.
-
Updated
Jun 10, 2021 - Python
-
Updated
Jun 24, 2021 - C++
Describe the bug
When compiling v0.13.0 build fails with
/home/zannzetsu/open3d/src/open3d/cpp/open3d/visualization/rendering/filament/FilamentEntitiesMods.cpp:212:22: required from here
/usr/include/fmt/core.h:1579:7: error: static assertion failed: Cannot format an argument. To make type T formattable provide a formatter<T> specialization: https://fmt.dev/latest/api.html#udt
157
-
Updated
Jul 18, 2021 - Go
For feature engineering tasks, I'd like to be able to determine whether a datetime is the beginning or end of a year, like I can in pandas.
import pandas as pd
s = pd.Series(["2021-02-27", "2020-03-31"], dtype="datetime64[ms]")
s.dt.is_year_end
0 False
1 False
dtype: bool
import pandas as pd
s = pd.Series(["2021-01-01", "2020-04-01"], dtype="datet
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
We're seeing a lot of warnings on Linux, host compiler GCC 9.3.0 with -Wconversion
.
For a basic example, compile:
#include <thrust/device_vector.h>
int main() {
thrust::device_vector<int> a;
return 0;
}
With: nvcc main.cu -Xcompiler=-Wconversion
.
This will result in around 1000 lines of output warnings.
Tested with the Thrust versions in CUDA 11.0, 11.3, and the la
请问可以直接training tmfile出来吗? 因为tengine-convert-tool covert 会有error
tengine-lite library version: 1.4-dev
Get input tensor failed
或是有例子能training出下面tmfile 呢?
 would serve the same purpose and be more concise.
We can just factor out the code from futhark bench
for this.
-
Updated
Jul 16, 2021 - Python
-
Updated
Jul 19, 2021 - Cuda
-
Updated
Feb 10, 2021 - C++
-
Updated
Dec 15, 2020 - Jupyter Notebook
-
Updated
Jul 13, 2021 - Python
-
Updated
Jun 30, 2021 - C
Created by Nvidia
Released June 23, 2007
- Website
- developer.nvidia.com/cuda-zone
- Wikipedia
- Wikipedia
Reporting a bug
visible in the change log (https://github.com/numba/numba/blob/master/CHANGE_LOG).
i.e. it's possible to run as 'python bug.py'.
(a static analyzer bug report)
In file
numba/np/ufunc/omppool.cpp
, `numba/np