PyTorch

PyTorch is an open source machine learning library based on the Torch library, used for applications such as computer vision and natural language processing, primarily developed by Facebook's AI Research lab.
Here are 20,072 public repositories matching this topic...
-
Updated
Jan 31, 2022 - Python
-
Updated
Feb 7, 2022 - Jupyter Notebook
-
Updated
Oct 16, 2021 - Python
-
Updated
Feb 9, 2022 - Python
-
Updated
Feb 10, 2022 - Jupyter Notebook
-
Updated
Feb 10, 2022 - JavaScript
I want to train a detector based on object365 dataset, but object365 is pretty large, and caused out of memory error in my computer.
I want to split the annotation file to 10, such as ann1,ann2,...ann10, then build 10 datasets and concatenate them, but I'm not sure whether it's
gonna work or not.
Any better suggestion?
-
Updated
Feb 9, 2022 - JavaScript
🐛 Bug
If apply_to_collections
is called on two dataclasses, the passed function
will be called on only one of the input dataclasses.
Dataclass inputs for apply_to_collection
are handled by a conditional:
apply_to_collections
does
-
Updated
Jan 19, 2022 - Python
-
Updated
Feb 10, 2022 - Python
-
Updated
Feb 2, 2022 - Python
-
Updated
Feb 7, 2022 - Jupyter Notebook
-
Updated
Feb 10, 2022 - Python
Change tensor.data
to tensor.detach()
due to
pytorch/pytorch#6990 (comment)
tensor.detach()
is more robust than tensor.data
.
🚀 Feature
Motivation
paper "LEARNING TO REPRESENT PROGRAMS WITH GRAPHS" which encode computer programs as graphs, with rich semantic information, however, most code implementation on this dataset VarMisuse is based on TensorFlow, like [tf-gnn-samples](https://github.com/microsof
-
Updated
Jan 20, 2022 - Python
-
Updated
Feb 10, 2022 - C++
-
Updated
Dec 30, 2021
Although the results look nice and ideal in all TensorFlow plots and are consistent across all frameworks, there is a small difference (more of a consistency issue). The result training loss/accuracy plots look like they are sampling on a lesser number of points. It looks more straight and smooth and less wiggly as compared to PyTorch or MXNet.
It can be clearly seen in chapter 6([CNN Lenet](ht
-
Updated
Aug 30, 2021 - Jupyter Notebook
-
Updated
Feb 7, 2022 - Python
Is your feature request related to a problem? Please describe.
I am uploading our dataset and models for the "Constructing interval measures" method we've developed, which uses item response theory to convert multiple discrete labels into a continuous spectrum for hate speech. Once we have this outcome our NLP models conduct regression rather than classification, so binary metrics are not r
New Operator
Describe the operator
Why is this operator necessary? What does it accomplish?
This is a frequently used operator in tensorflow/keras
Can this operator be constructed using existing onnx operators?
If so, why not add it as a function?
I don't know.
Is this operator used by any model currently? Which one?
Are you willing to contribute it?
-
Updated
Feb 9, 2022 - Python
-
Updated
Feb 9, 2022 - Python
-
Updated
Feb 10, 2022 - Python
Is your feature request related to a problem? Please describe.
I typically used compressed datasets (e.g. gzipped) to save disk space. This works fine with AllenNLP during training because I can write my dataset reader to load the compressed data. However, the predict
command opens the file and reads lines for the Predictor
. This fails when it tries to load data from my compressed files.
-
Updated
Jul 25, 2021 - Jupyter Notebook
Created by Facebook's AI Research lab (FAIR)
Released September 2016
Latest release about 2 months ago
- Repository
- pytorch/pytorch
- Website
- pytorch.org
- Wikipedia
- Wikipedia
Related to #5142,
AlbertTokenizer
(which uses SentencePiece) doesn't decode special tokens (like [CLS], [MASK]) properly. This issue was discovered when adding the Nystromformer model (#14659), which uses this tokenizer.To reproduce (Transformers v4.15 or below):