gpu
Here are 1,904 public repositories matching this topic...
-
Updated
Nov 12, 2020 - Jupyter Notebook
-
Updated
Sep 16, 2020 - Makefile
At this moment relu_layer op doesn't allow threshold configuration, and legacy RELU op allows that.
We should add configuration option to relu_layer.
-
Updated
Oct 7, 2020 - JavaScript
-
Updated
Nov 12, 2020 - Python
Problem:
catboost version: 0.23.2
Operating System: all
Tutorial: https://github.com/catboost/tutorials/blob/master/custom_loss/custom_metric_tutorial.md
Impossible to use custom metric (С++).
Code example
from catboost import CatBoost
train_data = [[1, 4, 5, 6],
-
Updated
Nov 4, 2020 - Python
-
Updated
Oct 29, 2020 - Python
-
Updated
Nov 12, 2020 - Jupyter Notebook
-
Updated
Nov 6, 2020 - Python
-
Updated
Apr 24, 2020 - Jsonnet
-
Updated
Nov 12, 2020 - C++
-
Updated
Jun 13, 2020 - HTML
Hi ,
I have tried out both loss.backward() and model_engine.backward(loss) for my code. There are several subtle differences that I have observed , for one retain_graph = True does not work for model_engine.backward(loss) . This is creating a problem since buffers are not being retained every time I run the code for some reason.
Please look into this if you could.
Current default value for rows_per_chunk
parameter of the CSV writer is 8, which means that the input table is by default broken into many small slices that are written out sequentially. This reduces the performance by an order on magnitude in some cases.
In Python layer, the default is the number of rows (i.e. write table out in a single pass). We can follow this by setting rows_per_chunk
Current implementation of join can be improved by performing the operation in a single call to the backend kernel instead of multiple calls.
This is a fairly easy kernel and may be a good issue for someone getting to know CUDA/ArrayFire internals. Ping me if you want additional info.
We would like to forward a particular 'key' column which is part of the features to appear alongside the predictions - this is to be able to identify to which set of features a particular prediction belongs to. Here is an example of predictions output using the tensorflow.contrib.estimator.multi_class_head:
{"classes": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"],
"scores": [0.068196
PR NVIDIA/cub#218 fixes this CUB's radix sort. We should:
- Check whether Thrust's other backends handle this case correctly.
- Provide a guarantee of this in the stable_sort documentation.
- Add regression tests to enforce this on all backends.
-
Updated
Nov 12, 2020 - C++
-
Updated
Oct 29, 2020 - CMake
-
Updated
Nov 12, 2020 - Jupyter Notebook
Improve this page
Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."
__init__
onSparseAdam
iterates overparams
to check if sparseTensor
s are included, but it does not take a possibility that the params is actually a generator for example in a case where we usemodel.parameters
. Therefore it fails the initialization withValueError: optimizer got an empty parameter list
in such cases.To Reproduce
Steps to reproduce the behavior: