Skip to content
#

gpu

Here are 2,267 public repositories matching this topic...

nmichlo
nmichlo commented Jun 23, 2021

🚀 Feature

Add drop_remainder: bool = True support to torch.chunk and torch.split

  • similar in function to drop_last in torch.utils.data.DataLoader
  • If the length of the dimension is not perfectly divisible, drop the remaining elements from the returned arrays.

Add redistribute: bool = True support to torch.chunk and torch.split

  • Spread the elements evenly across the c
pseudotensor
pseudotensor commented Jan 12, 2021

Problem: the approximate method can still be slow for many trees
catboost version: master
Operating System: ubuntu 18.04
CPU: i9
GPU: RTX2080

Would be good to be able to specify how many trees to use for shapley. The model.predict and prediction_type versions allow this. lgbm/xgb allow this.

H2O is an Open Source, Distributed, Fast & Scalable Machine Learning Platform: Deep Learning, Gradient Boosting (GBM) & XGBoost, Random Forest, Generalized Linear Modeling (GLM with Elastic Net), K-Means, PCA, Generalized Additive Models (GAM), RuleFit, Support Vector Machine (SVM), Stacked Ensembles, Automatic Machine Learning (AutoML), etc.

  • Updated Jul 4, 2021
  • Jupyter Notebook
rsn870
rsn870 commented Aug 21, 2020

Hi ,

I have tried out both loss.backward() and model_engine.backward(loss) for my code. There are several subtle differences that I have observed , for one retain_graph = True does not work for model_engine.backward(loss) . This is creating a problem since buffers are not being retained every time I run the code for some reason.

Please look into this if you could.

solardiz
solardiz commented Jul 19, 2019

Our users are often confused by the output from programs such as zip2john sometimes being very large (multi-gigabyte). Maybe we should identify and enhance these programs to output a message to stderr to explain to users that it's normal for the output to be very large - maybe always or maybe only when the output size is above a threshold (e.g., 1 million bytes?)

esnvidia
esnvidia commented Jun 15, 2021

Describe the bug
Clipping a DataFrame or Series using ints causes a cudf Failure because it won't handle the different dtypes (int and float)

Steps/Code to reproduce bug

data = cudf.Series([-0.43, 0.1234, 1.5, -1.31])
data.clip(0, 1)

...
  File "cudf/_lib/replace.pyx", line 216, in cudf._lib.replace.clip
  File "cudf/_lib/replace.pyx", line 198, in cudf._lib.replace.clamp
pyaf
pyaf commented May 24, 2021

Describe the Problem

plot_model currently has the save argument which can be used to save the plots. It does not provide the functionality to decide where to save the plot and with what name. Right now it saves the plot with predefined names in the current working directory.

Describe the solution you'd like

We can have another argument save_path which is used whenever the `

Improve this page

Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."

Learn more