gpu
Here are 1,858 public repositories matching this topic...
-
Updated
Oct 6, 2020 - Jupyter Notebook
-
Updated
Sep 16, 2020 - Makefile
At this moment relu_layer op doesn't allow threshold configuration, and legacy RELU op allows that.
We should add configuration option to relu_layer.
-
Updated
May 29, 2020 - JavaScript
-
Updated
Oct 6, 2020 - Python
-
Updated
Aug 17, 2020 - Python
Problem:
catboost version: 0.23.2
Operating System: all
Tutorial: https://github.com/catboost/tutorials/blob/master/custom_loss/custom_metric_tutorial.md
Impossible to use custom metric (С++).
Code example
from catboost import CatBoost
train_data = [[1, 4, 5, 6],
-
Updated
Oct 5, 2020 - Python
-
Updated
Oct 7, 2020 - Jupyter Notebook
-
Updated
Sep 17, 2020 - Python
-
Updated
Apr 24, 2020 - Jsonnet
-
Updated
Jun 13, 2020 - HTML
-
Updated
Oct 6, 2020 - C++
Hi ,
I have tried out both loss.backward() and model_engine.backward(loss) for my code. There are several subtle differences that I have observed , for one retain_graph = True does not work for model_engine.backward(loss) . This is creating a problem since buffers are not being retained every time I run the code for some reason.
Please look into this if you could.
Improve readability of thread id based branches by giving them more descriptive names.
e.g.
if (!t) // is actually a t == 0
and
https://github.com/rapidsai/cudf/blob/57ef76927373d7260b6a0eda781e59a4c563d36e/cpp/src/io/statistics/column_stats.cu#L285
Is actually a lane_id == 0
As demonstrated in rapidsai/cudf#6241 (comment), pr
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
We would like to forward a particular 'key' column which is part of the features to appear alongside the predictions - this is to be able to identify to which set of features a particular prediction belongs to. Here is an example of predictions output using the tensorflow.contrib.estimator.multi_class_head:
{"classes": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"],
"scores": [0.068196
Would it make sense to add a variadic overload of make_zip_iterator
that composes the existing make_zip_iterator
with make_tuple
? I have this in my own code, and I find that it reduces syntactic overhead.
template<typename... Iterators>
__host__ __device__
zip_iterator<thrust::tuple<Iterators...>> make_zip_iterator(thrust::tuple<Iterators...> t)
{
return zip_iterator<thrust::tupl
-
Updated
Oct 7, 2020 - C++
-
Updated
Sep 25, 2020 - CMake
Hey everyone!
mapd-core-cpu is already available on conda-forge (https://anaconda.org/conda-forge/omniscidb-cpu)
now we should add some instructions on the documentation.
at this moment it is available for linux and osx.
some additional information about the configuration:
- for now, always install
omniscidb-cpu
inside a conda environment (also it is a good practice), eg:
Improve this page
Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."
Compiling against the C++ API on macOS using GCC-9.3, and cmake seems to use a bad flag:
... -fopenmp -D_GLIBCXX_USE_CXX11_ABI= -std=c++14 ...
-- note how it "blanks out" the_GLIBCXX_USE_CXX11_ABI
variable. This causes the compiler to fail in the stdlib: