Skip to content
#

apache-spark

Here are 900 public repositories matching this topic...

yeikel
yeikel commented Jan 4, 2019

According to the generated build

The commands to launch are the following :

docker pull andypetrella/spark-notebook:0.7.0-scala-2.11.8-spark-2.1.1-hadoop-2.7.2-with-hive
docker run -p 9001:9001 andypetrella/spark-notebook:0.7.0-scala-2.11.8-spark-2.1.1-hadoop-2.7.2-with-hive

Using that image (and I think it i

shane-huang
shane-huang commented Apr 29, 2020
imback82
imback82 commented Mar 19, 2019

We create multiple jars during our builds to accommodate multiple versions of Apache Spark. In the current approach, the implementation is copied from one version to another and then necessary changes are made.

An ideal approach could create a common directory and extract common classes from duplicate code. Note that even if class/code is exactly the same, you cannot pull out to a common clas

andrewjw1995
andrewjw1995 commented May 16, 2018

The documentation file appears to have been generated with no space between the hashes and the header text. This is causing the headers to not display correctly, and is difficult to read. See below for an example of with and without the space:

##

Mobius API Documentation


###Microsoft.Spark.CSharp.Core.Accumulator</

sparklyr
scworland
scworland commented Jan 23, 2019

The "components" returned from ml_pca() are NULL

# example
sc <- spark_connect(master = "local")
iris_tbl <- sdf_copy_to(sc, iris, name = "iris_tbl", overwrite = TRUE)

pca <- iris_tbl %>%
  select(-Species) %>%
  ml_pca()

pca$components
NULL

R session information:

devtools::session_info()
Session info -----------------------------------------------------------
facundominguez
facundominguez commented Dec 19, 2016

While trying to write some tests of sparkle using tasty I found that it doesn't seem to work when bound threads other than the main one are used. The following program fails with:

$ stack --nix exec -- spark-submit --master 'local[1]' sparkle-example-osthreads.jar
16/12/19 10:30:29 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes 

Improve this page

Add a description, image, and links to the apache-spark topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the apache-spark topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.