-
Updated
May 18, 2020
text-mining
Here are 1,105 public repositories matching this topic...
-
Updated
Oct 7, 2019 - Python
-
Updated
Apr 18, 2018 - R
Hi there,
I think there might be a mistake in the documentation. The Understanding Scaled F-Score
section says
The F-Score of these two values is defined as:
$$ \mathcal{F}_\beta(\mbox{prec}, \mbox{freq}) = (1 + \beta^2) \frac{\mbox{prec} \cdot \mbox{freq}}{\beta^2 \cdot \mbox{prec} + \mbox{freq}}. $$
$\beta \in \mathcal{R}^+$ is a scaling factor where frequency is favored if $\beta
-
Updated
Jan 22, 2020
$ make show_docs
or
$ cd docs && make html
or
$ cd docs && sphinx-build -v -b html -d _build/doctrees . _build/html
Sphinx 버전 2.2.1 실행 중
Traceback (most recent call last):
File "/Users/minhoryang/.anyenv/envs/pyenv/versions/3.7.4-konlpy/lib/python3.7/site-packages/sphinx/cmd/build.py", line 275, in build_main
args.tags, args.verbosity, args.jobs, args.keep_
-
Updated
Aug 21, 2018 - Java
In the current tidytext
document explaining about the tidy approach to stm
object, there is no specific example of how to add covariates.
I wanted to try that out with stm::gadarian data using prevalence = ~treatment + s(pid_rep)
covariate formula; however, I have fac
-
Updated
Mar 7, 2020 - TeX
Hello,
I am getting the following error message "error: package directory 'rake_nltk' does not exist" when installing rake-nltk with:
git clone https://github.com/csurfer/rake-nltk.git
python rake-nltk/setup.py install :
I also tried the option pip install rake-nltk but the installation also fails:
File "/tmp/pip-build-2zTHYP/rake-nltk/setup.py", line 17, in _post_install
import
I would like to know what all the abbreviations mean? Some I can guess, like "PUNCT", but no idea what "X" might be. I want to retain contractions, but hard to choose options without documentation.
Thanks. Great performance code!
-
Updated
May 5, 2020 - C++
-
Updated
Apr 22, 2020 - Jupyter Notebook
When using artm.SmoothSparseThetaRegularizer(tau=tau_val) with tau_val<0 we get some \Theta matrix columns filled totally with zeros. From perplexity score, the optimization converges. The quantity of documents with all zeros in their \Theta columns grows as $tau_val->-\infty$.
How it's possible that optimization constraint on theta columns violates?
Hi,
I used to have a previous version of LDAvis (2014) installed with devtools.
In the version I had of LDAvis I would call createJSON as:
json <- createJSON(K, phi, term.frequency, vocab, topic.proportions)
Today I updated my R packages and have a newer vesion of LDAvis (from CRAN) which uses createJSON as:
json <- createJSON(phi, theta, doc.length, vocab, term.frequency)
I'm using MALLET for t
-
Updated
Sep 2, 2018
-
Updated
Nov 1, 2019 - Jupyter Notebook
-
Updated
May 13, 2020 - Python
-
Updated
Feb 26, 2018 - Python
-
Updated
May 15, 2020 - Python
I think it is necessary to add an experiment that compare the test accuracy of the original text and the adversarial text examples in the target model to judge whether the adversarial text examples really reduce the accuracy.
-
Updated
May 4, 2020
Ontology example
is someone familiar with the Ontology process and can share an RDF file for example?
Super Thanks :)
-
Updated
Feb 17, 2020 - Jupyter Notebook
-
Updated
Mar 30, 2020 - Jupyter Notebook
-
Updated
Dec 28, 2017 - Python
-
Updated
Oct 7, 2019 - Jupyter Notebook
Improve this page
Add a description, image, and links to the text-mining topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the text-mining topic, visit your repo's landing page and select "manage topics."
We're undergoing an internal software audit and identified at least one textract component released under the Affero GPL: the EbookLib.
Lawyers are getting a bit antsy over this. In general, compatibility with GPL means that code released under a different license (e.g. MIT) and combined with GPL'd code must be released under GPL. This might create a b