bigdata
Here are 1,219 public repositories matching this topic...
-
Updated
May 21, 2020
Hi,
I am trying to load a CSV with no header using
df = vaex.open('data/star0000-1.csv',sep=",", header=None, error_bad_lines=False)
but I get
could not convert column 0, error: TypeError('getattr(): attribute name must be string'), will try to convert it to string
Giving up column 0, error: TypeError('getattr(): attribute name must be string')
could not convert column
We create multiple jars during our builds to accommodate multiple versions of Apache Spark. In the current approach, the implementation is copied from one version to another and then necessary changes are made.
An ideal approach could create a common
directory and extract common classes from duplicate code. Note that even if class/code is exactly the same, you cannot pull out to a common clas
-
Updated
May 29, 2020 - Java
-
Updated
Sep 6, 2017 - Jupyter Notebook
On home page of website: https://nlp.johnsnowlabs.com/ I read "Full Python, Scala, and Java support"
Unfortunately it's 3 days now I'm trying to use Spark NLP in Java without any success.
- I cannot find Java API (JavaDoc) of the framework.
- not event a single example in Java is available
- I do not know Scala, I do not know how to convert things like:
val testData = spark.createDataFrame(
目前MLSQL已经提供了一份中文文档,
英文文档也有但很不完整同时不少内容过时了。现在诚邀各位英文较好的朋友
将中文文档翻译成英文。
参与翻译流程
- fork 项目 https://github.com/allwefantasy/streamingpro
- clone自己刚刚fork的项目
- 中文文档在 ·项目根目录/docs/gitbook/zh/
- 英文文档在 ·项目根目录/docs/gitbook/en/
- 申领中文文档下的文件
- 翻译稿件放在英文文档目录中
- push自己的成果,然后在web上选择 新提交PR
英文要和中文保持一致。领取任务者在本Issue下罗列领取的中文章节,避免其他人重复。
譬如 我是用户A, 我可以根据目录文件进行
-
Updated
May 30, 2020 - Java
Is this a BUG REPORT or FEATURE REQUEST?:
/kind feature
What happened:
Very recently, Amazon announced [windows support](https://aws
The documentation file appears to have been generated with no space between the hashes and the header text. This is causing the headers to not display correctly, and is difficult to read. See below for an example of with and without the space:
##
Mobius API Documentation
###Microsoft.Spark.CSharp.Core.Accumulator</
Because some user has had problems configuring these services could be helpful to make some examples or videos about how to properly setup Optimus in this services.
-
Updated
May 16, 2020 - Python
-
Updated
May 11, 2020 - Scala
-
Updated
Dec 24, 2019 - Jupyter Notebook
-
Updated
Apr 14, 2020 - Go
Description
Customer have unique requirements for alarming, specifically with respect to log processing.
As a customer, I would like to know how to configure alarms on log processing volume.
Acceptance Criteria
- Document in operations doc page how to add alarms on log processing with examples
-
Updated
Apr 24, 2020 - Python
-
Updated
Feb 15, 2020
When using artm.SmoothSparseThetaRegularizer(tau=tau_val) with tau_val<0 we get some \Theta matrix columns filled totally with zeros. From perplexity score, the optimization converges. The quantity of documents with all zeros in their \Theta columns grows as $tau_val->-\infty$.
How it's possible that optimization constraint on theta columns violates?
The installation section currently says to use npm install --save circos
with no further instruction.
Doing just this and then using Browserify to require circos results in the following error:
Error: Cannot find module '!!../../node_modules/css-loader/index.js!./tooltip.css' from '/.../node_modules/circos/dist'
Is there an assumption that Webpack is being used by those who try to impo
As outlined in #16, it's often useful to extend fine-grained control of sharding to the user. It can be solved by wrapping integers with an identity hash function, but that seems less than ideal. It might be useful to provide this functionality as part of bigslice.Reshuffle
.
Improve this page
Add a description, image, and links to the bigdata topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the bigdata topic, visit your repo's landing page and select "manage topics."
必须先将通过STable创建的表先删除,然后才能删除超级表。当级联表很多时就麻烦了,除非写代码来删除。