Apache Spark

Apache Spark is an open source distributed general-purpose cluster-computing framework. It provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
Here are 6,778 public repositories matching this topic...
-
Updated
Apr 3, 2022 - Python
-
Updated
Jul 7, 2022 - Python
-
Updated
Jul 1, 2022 - Go
-
Updated
May 18, 2022
-
Updated
Jul 7, 2022 - Python
At this moment relu_layer op doesn't allow threshold configuration, and legacy RELU op allows that.
We should add configuration option to relu_layer.
-
Updated
Jun 21, 2022 - Java
-
Updated
Jun 9, 2022
-
Updated
Jun 17, 2022 - Java
-
Updated
Jul 7, 2022 - Jupyter Notebook
-
Updated
Feb 8, 2022 - Python
-
Updated
Jul 7, 2022 - Java
Feature request
Overview
Currently, the DELETE
operation returns an empty result. It would be more useful if it returned the number of deleted rows.
Motivation
The number of deleted rows is an obvious metric that users would want from a delete operation.
Further details
Currently, DeleteCommand.scala
is explicitly returning an empty DataFrame
[here](https://g
-
Updated
Apr 24, 2020 - Jsonnet
Problem:
The current log will output something like val_function_0
while it should be val_mean_squared_error_0
.
Solution:
"val/{}_{}".format(type(metric).__name__, i)
use the name of the type of metric
(metric is an instance of torchmetrics.metric.Metric
so the type of it is function
), that's why the output looks like val_function_0
. It should use the name of the metri
-
Updated
Apr 21, 2022 - Python
I have a simple regression task (using a LightGBMRegressor) where I want to penalize negative predictions more than positive ones. Is there a way to achieve this with the default regression LightGBM objectives (see https://lightgbm.readthedocs.io/en/latest/Parameters.html)? If not, is it somehow possible to define (many example for default LightGBM model) and pass a custom regression objective?
-
Updated
May 18, 2022 - Scala
-
Updated
May 12, 2021 - Jupyter Notebook
-
Updated
Oct 19, 2021 - JavaScript
-
Updated
Jul 7, 2022 - Scala
Used Spark version
Spark Version: 2.4.4
Used Spark Job Server version
SJS version: v0.11.1
Deployed mode
client on Spark Standalone
Actual (wrong) behavior
I can't get config, when post a job with 'sync=true'. I got it:
http://localhost:8090/jobs/ff99479b-e59c-4215-b17d-4058f8d97d25/config
{"status":"ERROR","result":"No such job ID ff99479b-e59c-4215-b17d-4058f8d97d25"
Created by Matei Zaharia
Released May 26, 2014
- Repository
- apache/spark
- Website
- spark.apache.org
- Wikipedia
- Wikipedia
Describe the bug
Using a time dimension on a runningTotal measure on Snowflake mixes quoted and unquoted columns in the query. This fails the query, because Snowflake has specific rules about quoted columns. Specifically:
So "date_from" <> date_from
To Reproduce
Steps to reproduce