Skip to content
#

numpy

Here are 6,928 public repositories matching this topic...

Data science Python notebooks: Deep learning (TensorFlow, Theano, Caffe, Keras), scikit-learn, Kaggle, big data (Spark, Hadoop MapReduce, HDFS), matplotlib, pandas, NumPy, SciPy, Python essentials, AWS, and various command lines.

  • Updated Oct 1, 2020
  • Python
numba
gmarkall
gmarkall commented Nov 3, 2020

PR #6447 adds a public API to get the maximum number of registers per thread (numba.cuda.Dispatcher.get_regs_per_thread()). There are other attributes that might be nice to provide - shared memory per block, local memory per thread, const memory usage, maximum block size.

These are all available in the FuncAttr named tuple: https://github.com/numba/numba/blob/master/numba/cuda/cudadrv/drive

人工智能学习路线图,整理近200个实战案例与项目,免费提供配套教材,零基础入门,就业实战!包括:Python,数学,机器学习,数据分析,深度学习,计算机视觉,自然语言处理,PyTorch tensorflow machine-learning,deep-learning data-analysis data-mining mathematics data-science artificial-intelligence python tensorflow tensorflow2 caffe keras pytorch algorithm numpy pandas matplotlib seaborn nlp cv等热门领域

  • Updated Feb 6, 2020
gluon-nlp
sxjscience
sxjscience commented Nov 19, 2020

Description

We should use the official mxnet batchify functions to implement our own batchify functions. However, since we'd like to later support other frameworks, we should still keep our own batchify.py. We can change it to call MXNet implementations.

MXNet batchify: https://github.com/apache/incubator-mxnet/blob/master/python/mxnet/gluon/data/batchify.py
GluonNLP batchify: https://gi

Improve this page

Add a description, image, and links to the numpy topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the numpy topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.