-
Updated
May 19, 2020 - Jupyter Notebook
exploratory-data-analysis
Here are 941 public repositories matching this topic...
Hi there,
I think there might be a mistake in the documentation. The Understanding Scaled F-Score
section says
The F-Score of these two values is defined as:
$$ \mathcal{F}_\beta(\mbox{prec}, \mbox{freq}) = (1 + \beta^2) \frac{\mbox{prec} \cdot \mbox{freq}}{\beta^2 \cdot \mbox{prec} + \mbox{freq}}. $$
$\beta \in \mathcal{R}^+$ is a scaling factor where frequency is favored if $\beta
-
Updated
Sep 9, 2018 - Jupyter Notebook
-
Updated
Nov 2, 2016 - OpenEdge ABL
-
Updated
Dec 20, 2019 - R
-
Updated
May 14, 2018 - Jupyter Notebook
-
Updated
Oct 22, 2018 - Java
Hello,
First of all, thanks for the great package.
I'm trying to compute density maps of a 3 dimensional points distribution. I understood from the documentation that a variable bandwith method was available but I couldn't figure out how to set up this option.
Additionnaly, in the case of a fixed bandwidth KDE for multidimensional data, I would have expected as in the stats_models_multivari
To improve spotting differences between datasets visually
(especially when there are many columns) it would be helpful if one could sort the categorical columns by the Jensen–Shannon divergence.
The code below tries to do so but it seems to distort the labels on the y-axis. Also, in case the jsd
column contains missing values, those variables are deleted from the graph.
library(in
-
Updated
Oct 31, 2019 - Jupyter Notebook
-
Updated
May 14, 2020 - HTML
-
Updated
May 19, 2019 - Jupyter Notebook
-
Updated
Feb 18, 2019 - Jupyter Notebook
-
Updated
Sep 6, 2017 - Jupyter Notebook
-
Updated
Jan 20, 2018 - R
-
Updated
Apr 7, 2018 - Jupyter Notebook
-
Updated
Mar 4, 2019 - Python
-
Updated
Jul 8, 2018 - Python
-
Updated
Apr 18, 2018 - HTML
Adding a description for the parameters will help the users understand how to specify values for each parameter. For example, the format of the longitude in Yelp.businesses table; the maximum limit of the results that a user can expect (if we incorporate limit parameter in the future).
-
Updated
Mar 22, 2020 - R
-
Updated
Apr 23, 2019 - Jupyter Notebook
-
Updated
Feb 6, 2018 - R
-
Updated
Jan 2, 2019 - Jupyter Notebook
-
Updated
May 19, 2020 - Python
Great and
very clear stepXstep package tutorial, Matt!.
A time-saving suggestion (if I may):
in Step:
"Examining the Results" (after Step 3),
where you have:
marketing_campaign_correlated_tbl %>%
filter(feature %in% c("DURATION", "POUTCOME", "PDAYS",
"PREVIOUS", "CONTACT", "HOUSING")) %>%
plot_correlation_funnel(interactive = FALSE, limits = c(-0.4, 0.4))
Why not "automatica
-
Updated
Jul 16, 2019 - R
-
Updated
Jun 15, 2018 - Python
-
Updated
May 31, 2017 - HTML
Improve this page
Add a description, image, and links to the exploratory-data-analysis topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the exploratory-data-analysis topic, visit your repo's landing page and select "manage topics."
As a user,
It would be nice to have the "Observed Value" Field be standardized to show percentages of "successful" validations, vs a mix of 0% / 100%. This causes confusion as there are different levels of validation outputs with different verbage (making someone not used to the expectations confused) I've given an example below in a screenshot for what I mean:
![image](https://user-images.g