Are you over 18 and want to see adult content?
More Annotations
#1 in Ayahuasca Education | Best Ayahuasca Retreat Peru/USA
Are you over 18 and want to see adult content?
Educational CyberPlayGround ® K-12 Education reference directory of Resources.
Are you over 18 and want to see adult content?
Tamil Cinema News | Tamil News | Sports News | Latest Tamil Movie News
Are you over 18 and want to see adult content?
Software online para procesar imágenes de drone - Aerial Insights
Are you over 18 and want to see adult content?
Kafélanka - objevování zapomenutých míst
Are you over 18 and want to see adult content?
Borderline Personality Disorder | BPDFamily.com
Are you over 18 and want to see adult content?
PremiaPão Publicidade em Sacos de Pão | A Maior Microfranquia do Brasil
Are you over 18 and want to see adult content?
Ihr Schnäppchen-Portal für Multimedia, Haushalt, Freizeit & Mode | www.living24.de
Are you over 18 and want to see adult content?
Favourite Annotations
A complete backup of https://townsvillenorthqueensland.com.au
Are you over 18 and want to see adult content?
A complete backup of https://geodesic-tents.com
Are you over 18 and want to see adult content?
A complete backup of https://robscholtemuseum.nl
Are you over 18 and want to see adult content?
A complete backup of https://sierraseybold.com
Are you over 18 and want to see adult content?
A complete backup of https://helpmepickit.com
Are you over 18 and want to see adult content?
A complete backup of https://darkthrone.com
Are you over 18 and want to see adult content?
A complete backup of https://mfj.gr.jp
Are you over 18 and want to see adult content?
A complete backup of https://joyprint.ru
Are you over 18 and want to see adult content?
A complete backup of https://point-s.ca
Are you over 18 and want to see adult content?
A complete backup of https://gerberchildrenswear.com
Are you over 18 and want to see adult content?
Text
SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README. SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK-CASSANDRA-CONNECTOR_2.11-1.3.1: 1.3.0 TO 1.4.0 Binary compatibility report for the spark-cassandra-connector_2.11-1.3.1 library between 1.3.0 and 1.4.0versions
SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README. SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK-CASSANDRA-CONNECTOR_2.11-1.3.1: 1.3.0 TO 1.4.0 Binary compatibility report for the spark-cassandra-connector_2.11-1.3.1 library between 1.3.0 and 1.4.0versions
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK PACKAGES
killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
Optimus Optimus is the missing library for cleansing (cleaning and much more) and pre-processing data in a distributed fashion withApache Spark.
SPARK PACKAGES
sparkling-graph Large scale, distributed graph processing made easy! Load your graph from multiple formats and compute measures (but not only) @sparkling-graph / Latest release: 0.0.7 (2017-05-16) / BSD2-Clause / (5)
SPARK-UTILS
This project contains some basic utilities that can help setting up a Spark application project. The SparkRunnable and SparkApp together with the configuration framework provide for easy Spark application creation with configuration that can be managed through configurationfiles or
TENSORFRAMES
TensorFrames (TensorFlow on Spark Dataframes) lets you manipulate Spark's DataFrames with TensorFlow programs. This package provides Python bindings, a Scala DSL and a small runtime to express and run TensorFlow computation graphs.SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) GMM - SPARK PACKAGES GMM algorithm models the entire data set as a finite mixture of Gaussian distributions,each parameterized by a mean vector, a covariance matrix and a mixture weights. Here the probability of each point to belong to each cluster is computed along with the cluster statistics. This distributed implementation of GMM in pysparkestimates the
SPARK-LLAP
A library to load data into Spark SQL DataFrames from Hive using LLAP. @hortonworks-spark / (0) A library to load data into Spark SQL DataFrames from Hive using LLAP. It also contains Catalog/Context classes to enable querying of Hive tables without having to SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README. SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK-CASSANDRA-CONNECTOR_2.11-1.3.1: 1.3.0 TO 1.4.0 Binary compatibility report for the spark-cassandra-connector_2.11-1.3.1 library between 1.3.0 and 1.4.0versions
SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK PACKAGES
killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
Optimus Optimus is the missing library for cleansing (cleaning and much more) and pre-processing data in a distributed fashion withApache Spark.
SPARK PACKAGES
sparkling-graph Large scale, distributed graph processing made easy! Load your graph from multiple formats and compute measures (but not only) @sparkling-graph / Latest release: 0.0.7 (2017-05-16) / BSD2-Clause / (5)
SPARK-UTILS
This project contains some basic utilities that can help setting up a Spark application project. The SparkRunnable and SparkApp together with the configuration framework provide for easy Spark application creation with configuration that can be managed through configurationfiles or
TENSORFRAMES
TensorFrames (TensorFlow on Spark Dataframes) lets you manipulate Spark's DataFrames with TensorFlow programs. This package provides Python bindings, a Scala DSL and a small runtime to express and run TensorFlow computation graphs.SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) GMM - SPARK PACKAGES GMM algorithm models the entire data set as a finite mixture of Gaussian distributions,each parameterized by a mean vector, a covariance matrix and a mixture weights. Here the probability of each point to belong to each cluster is computed along with the cluster statistics. This distributed implementation of GMM in pysparkestimates the
SPARK-LLAP
A library to load data into Spark SQL DataFrames from Hive using LLAP. @hortonworks-spark / (0) A library to load data into Spark SQL DataFrames from Hive using LLAP. It also contains Catalog/Context classes to enable querying of Hive tables without having to SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README. SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK-CASSANDRA-CONNECTOR_2.11-1.3.1: 1.3.0 TO 1.4.0 Binary compatibility report for the spark-cassandra-connector_2.11-1.3.1 library between 1.3.0 and 1.4.0versions
SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README. SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK-CASSANDRA-CONNECTOR_2.11-1.3.1: 1.3.0 TO 1.4.0 Binary compatibility report for the spark-cassandra-connector_2.11-1.3.1 library between 1.3.0 and 1.4.0versions
SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
mezzanine Mezzanine is a library built on Spark Streaming used to consume data from Kafka and store it into Hadoop. @groupon / Latest release: 1.0 (2016-07-29) / BSD 3-Clause / (0)SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
sparkling-graph Large scale, distributed graph processing made easy! Load your graph from multiple formats and compute measures (but not only) @sparkling-graph / Latest release: 0.0.7 (2017-05-16) / BSD2-Clause / (5)
SPARK PACKAGES
Optimus Optimus is the missing library for cleansing (cleaning and much more) and pre-processing data in a distributed fashion withApache Spark.
SPARK-UTILS
This project contains some basic utilities that can help setting up a Spark application project. The SparkRunnable and SparkApp together with the configuration framework provide for easy Spark application creation with configuration that can be managed through configurationfiles or
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited)TENSORFRAMES
TensorFrames (TensorFlow on Spark Dataframes) lets you manipulate Spark's DataFrames with TensorFlow programs. This package provides Python bindings, a Scala DSL and a small runtime to express and run TensorFlow computation graphs.SPARK-XML-UTILS
Spark-xml-utils provides the ability to filter documents based on an xpath expression, return specific nodes for an xpath/xquery expression, or transform documents using a xslt stylesheet. The spark-xml-utils library was developed because there is a large amount of xml in some big datasets and I felt this data could be betterserved by
SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
mezzanine Mezzanine is a library built on Spark Streaming used to consume data from Kafka and store it into Hadoop. @groupon / Latest release: 1.0 (2016-07-29) / BSD 3-Clause / (0)SPARK PACKAGES
killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
sparkling-graph Large scale, distributed graph processing made easy! Load your graph from multiple formats and compute measures (but not only) @sparkling-graph / Latest release: 0.0.7 (2017-05-16) / BSD2-Clause / (5)
SPARK PACKAGES
Optimus Optimus is the missing library for cleansing (cleaning and much more) and pre-processing data in a distributed fashion withApache Spark.
SPARK-UTILS
This project contains some basic utilities that can help setting up a Spark application project. The SparkRunnable and SparkApp together with the configuration framework provide for easy Spark application creation with configuration that can be managed through configurationfiles or
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) GMM - SPARK PACKAGES GMM algorithm models the entire data set as a finite mixture of Gaussian distributions,each parameterized by a mean vector, a covariance matrix and a mixture weights. Here the probability of each point to belong to each cluster is computed along with the cluster statistics. This distributed implementation of GMM in pysparkestimates the
SPARK-XML-UTILS
Spark-xml-utils provides the ability to filter documents based on an xpath expression, return specific nodes for an xpath/xquery expression, or transform documents using a xslt stylesheet. The spark-xml-utils library was developed because there is a large amount of xml in some big datasets and I felt this data could be betterserved by
SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager. SPARK PACKAGESDIBBHATTWXHC3SC6OPM8M1HXBOMY killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
spark-solr Tools for reading data from Solr as a Spark RDD and indexing objects from Spark into Solr using SolrJ. @LucidWorks / Latest release: 2.0.1 (2016-06-09) / Apache-2.0 / (1)SPARK-SAS7BDAT
Splittable SAS (.sas7bdat) Input Format for Hadoop and Spark SQL. @saurfang / (1) This packages allow reading SAS binary file (.sas7bdat) in parallel as data frame in Spark SQL. It provides utility to export it as CSV (using spark-csv) or parquet file.SPARK-DARIA
When the Spark source code doesn't provide functionality, turn to thislibrary
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) SPARK-STREAMING-JDBC-SOURCE JDBC source for spark structured streaming. @sutugin / (1) A library for querying JDBC data with Apache Spark Structured Streaming, for Spark SQL and DataFrames.SPARK-DICOM
This package doesn't have any releases published in the Spark Packages repo, or with maven coordinates supplied. You may have to build this package from source, or it may simply be a script. To use this Spark Package, please follow the instructions in the README.SPARK-ON-HPC
Script to submit spark jobs on a traditional HPC cluster. Spark-on-HPC dynamically provisions Apache Spark clusters and run spark jobs on an HPC under its traditional resource manager.SPARK PACKAGES
zen Zen aims to provide the largest scale and the most efficient machine learning platform on top of Spark, including but not limited to logistic regression, latent dirichilet allocation, factorizationmachines and DNN.
SPARK PACKAGES
spark-crossdata SparkSQL extension as a library for Apache Spark extending and improving its capabilities for a data federation system. @Stratio / Latest release: 1.4.0 (2016-07-06) / Apache-2.0 / (6)SPARK PACKAGES
mezzanine Mezzanine is a library built on Spark Streaming used to consume data from Kafka and store it into Hadoop. @groupon / Latest release: 1.0 (2016-07-29) / BSD 3-Clause / (0)SPARK PACKAGES
killrweather KillrWeather is a reference application (in progress) showing how to easily leverage and integrate Apache Spark, Apache Cassandra, and Apache Kafka for fast, streaming computations on time series data in asynchronous Akka event-driven environments.SPARK PACKAGES
sparkling-graph Large scale, distributed graph processing made easy! Load your graph from multiple formats and compute measures (but not only) @sparkling-graph / Latest release: 0.0.7 (2017-05-16) / BSD2-Clause / (5)
SPARK PACKAGES
Optimus Optimus is the missing library for cleansing (cleaning and much more) and pre-processing data in a distributed fashion withApache Spark.
SPARK-UTILS
This project contains some basic utilities that can help setting up a Spark application project. The SparkRunnable and SparkApp together with the configuration framework provide for easy Spark application creation with configuration that can be managed through configurationfiles or
SPARK-SCDENGINE
Spark 1.6.0. Features: 1.Atleast 10x lesser time to implement (as compared to Informatica BDE implementation) 2.Faster performance (as compared to HIVE & Tez Queries) 3.Plug-n-Play application with simple configuration (just provide few details about source & target table) 4. Auto Datatype conversions (limited) GMM - SPARK PACKAGES GMM algorithm models the entire data set as a finite mixture of Gaussian distributions,each parameterized by a mean vector, a covariance matrix and a mixture weights. Here the probability of each point to belong to each cluster is computed along with the cluster statistics. This distributed implementation of GMM in pysparkestimates the
SPARK-XML-UTILS
Spark-xml-utils provides the ability to filter documents based on an xpath expression, return specific nodes for an xpath/xquery expression, or transform documents using a xslt stylesheet. The spark-xml-utils library was developed because there is a large amount of xml in some big datasets and I felt this data could be betterserved by
Toggle navigation
__
* Feedback
* Register a package* Login
A COMMUNITY INDEX OF THIRD-PARTY PACKAGES FOR APACHE SPARK.
SHOWING PACKAGES 1 - 50 OUT OF 500Next >
* All (500)
* Core (14)
* Data Sources (57)
* Machine Learning (101)* Streaming (63)
* Graph (21)
* PySpark (23)
* Applications (17)
* Deployment (12)
* Examples (26)
* Tools (36)
SPARK-ALS
ANOTHER, HOPEFULLY BETTER, IMPLEMENTATION OF ALS ON SPARK (ALREADYMERGED INTO MLLIB)
@mengxr / Latest release: 0.1.0 (2014-11-27) / BSD3-Clause / (__1)
* 3ml
* 2mllib
* 2recommendation
*
-------------------------MLLIB-GRID-SEARCH
AN EXAMPLE PROJECT FOR DOING GRID SEARCH IN MLLIB @spark-ml / Latest release: 0.0.1 (2014-11-27) / BSD3-Clause / (__2)
* 1ml
* 1example
* 1examples
*
-------------------------SPARK-AVRO
INTEGRATION UTILITIES FOR USING SPARK WITH APACHE AVRO DATA @databricks / Latest release: 4.0.0-s_2.11 (2017-10-30) / Apache-2.0 / (__13)* 6sql
* 4input
* 4avro
*
-------------------------SPARK-REDSHIFT
REDSHIFT DATA SOURCE FOR APACHE SPARK @databricks / Latest release: 3.0.0-preview1 (2016-11-01) / Apache-2.0 / (__3)* 2sql
* 2data source
* 2redshift
*
------------------------- KAFKA-SPARK-CONSUMER HIGH PERFORMANCE KAFKA CONSUMER FOR SPARK STREAMING.SUPPORTS MULTI TOPIC FETCH, KAFKA SECURITY. RELIABLE OFFSET MANAGEMENT IN ZOOKEEPER. NO DATA-LOSS. NO DEPENDENCY ON HDFS AND WAL. IN-BUILT PID RATE CONTROLLER. SUPPORT MESSAGE HANDLER . OFFSET LAG CHECKER @dibbhatt / Latest release: 2.1.0 (2019-08-28) /Apache-2.0 / (__7)
* 4streaming
* 3kafka
*
-------------------------THUNDER
LARGE-SCALE NEURAL DATA ANALYSIS WITH SPARK @freeman-lab / Latest release: 0.4.1 (2014-11-27) / BSD 3-Clause / (__6)* 3neuroscience
* 2python
* 2machine learning
*
-------------------------SPARK_AZURE
SPARK LAUNCH SCRIPT FOR MICROSOFT AZURE @sigmoidanalytics / No release yet / (__10)* 1Azure
* 1spark
* 1Microsoft
*
-------------------------SPORK
PIG ON APACHE SPARK
@sigmoidanalytics / No release yet / (__9)* 1streaming
* 1spark
* 1pig
*
-------------------------SPARK_GCE
SPARK LAUNCH SCRIPT FOR GOOGLE COMPUTE ENGINE @sigmoidanalytics / No release yet / (__8) * 1Google Compute Engine* 1spark
* 1GCE
*
-------------------------SPARK-JOBSERVER
REST JOB SERVER FOR SPARK @spark-jobserver / No release yet / (__3)* 1application
* 1REST
* 1Mesos
*
-------------------------GMM
GAUSSIAN MIXTURE MODEL IMPLEMENTATION IN PYSPARK @FlytxtRnD / Latest release: 0.1 (2015-04-07) /EPL-1.0 / (__5)
* 1python
* 1mllib
*
-------------------------SPARK-CSV
SPARK SQL CSV DATA SOURCE @databricks / Latest release: 1.5.0-s_2.11 (2016-09-07) / Apache-2.0 / (__10)* 4csv
* 3sql
* 2DataSource
*
-------------------------SPARK-INDEXEDRDD
AN EFFICIENT UPDATABLE KEY-VALUE STORE FOR APACHE SPARK @amplab / Latest release: 0.4.0 (2017-01-11) /Apache-2.0 / (__1)
* 2core
* 2kv
* 1anothertag
*
-------------------------SPARK-PERF
PERFORMANCE TESTS FOR SPARK @databricks / No release yet / (__1) -------------------------KILLRWEATHER
KILLRWEATHER IS A REFERENCE APPLICATION (IN PROGRESS) SHOWING HOW TO EASILY LEVERAGE AND INTEGRATE APACHE SPARK, APACHE CASSANDRA, AND APACHE KAFKA FOR FAST, STREAMING COMPUTATIONS ON TIME SERIES DATA IN ASYNCHRONOUS AKKA EVENT-DRIVEN ENVIRONMENTS. @killrweather / No release yet / (__1)* 1streaming
*
-------------------------SPARK-HASH
LOCALITY SENSITIVE HASHING FOR APACHE SPARK @mrsqueeze / No release yet / (__0)* 1mllib
* 1lsh
*
-------------------------ZEPPELIN
ZEPPELIN, A WEB-BASED NOTEBOOK THAT ENABLES INTERACTIVE DATAANALYTICS.
@NFLabs / No release yet / (__3)* 1Applications
* 1notebook
* 1interactive
*
-------------------------SPARK-HBASE
INTEGRATION UTILITIES FOR USING SPARK WITH APACHE HBASE DATA @haosdent / No release yet / (__1)* 1hbase
*
-------------------------SPARK-DBF
SPARK SQL DBF LIBRARY @mraad / No release yet / (__0)* 1sql
*
-------------------------SPARK_HBASE
THE EXAMPLE IN SCALA OF READING DATA SAVED IN HBASE BY SPARK AND THE EXAMPLE OF CONVERTER FOR PYTHON @GenTang / No release yet / (__3)* 1python
* 1hbase
*
-------------------------SPARKLING
A CLOJURE LIBRARY FOR APACHE SPARK: FAST, FULLY-FEATURES, ANDDEVELOPER FRIENDLY
@gorillalabs / Latest release: 1.0.0 (2014-12-31)/ EPL-1.0 / (__3)
* 2clojure
* 1API
*
-------------------------SPARK-KERNEL
A KERNEL THAT ENABLES APPLICATIONS TO INTERACT WITH APACHE SPARK. @ibm-et / No release yet / (__0)* 1ipython
* 1foundation
* 1interactive
*
-------------------------PYSPARK-PICTURES
LEARN THE PYSPARK API THROUGH PICTURES AND SIMPLE EXAMPLES @jkthompson / No release yet / (__0)* 2tutorials
* 2examples
*
-------------------------DEEP-SPARK
CONNECTING APACHE SPARK WITH DIFFERENT DATA STORES @Stratio / Latest release: 0.7.0-RC1 (2015-01-14) /Apache-2.0 / (__20)
* 6database
* 6mongo
* 6cassandra
*
------------------------- STREAMING-CEP-ENGINE STREAMING CEP ENGINE POWERED BY SPARK STREAMING & SIDDHI @Stratio / Latest release: 0.6.2 (2015-01-14) /Apache-2.0 / (__19)
* 5spark streaming
* 5cep
* 4complex event processing*
------------------------- GENERALIZED-KMEANS-CLUSTERING THIS PROJECT GENERALIZES THE SPARK MLLIB K-MEANS CLUSTERER TO SUPPORT ARBITRARY DISTANCE FUNCTIONS @derrickburns / No release yet / (__3)* 1clustering
* 1mllib
* 1machine learning
*
-------------------------SPARKLING-WATER
SPARKLING WATER PROVIDES H2O ALGORITHMS INSIDE SPARK CLUSTER @h2oai / Latest release: 1.4.3 (2015-07-06) / Apache-2.0/ (__2)
* 1h2o
* 1algorithms
* 1machine learning
*
-------------------------SPARK-ML-STREAMING
VISUALIZE STREAMING MACHINE LEARNING IN SPARK @freeman-lab / No release yet / (__1)* 1streaming
* 1machine learning
* 1visualization
*
-------------------------SPARK
APACHE CAMEL STREAMING CONSUMER @synsys / Latest release: 1.0.0 (2015-01-26) /Apache-2.0 / (__0)
* 1streaming
* 1consumer
* 1camel
*
------------------------- SPARK-HBASE-CONNECTOR CONNECT SPARK TO HBASE FOR READING AND WRITING DATA WITH EASE @nerdammer / Latest release: 1.0.3 (2016-04-20) /Apache-2.0 / (__3)
* 1streaming
* 1hbase
* 1library
*
-------------------------SPARK-TESTING-BASE
BASE CLASSES TO USE WHEN WRITING TESTS WITH SPARK @holdenk / Latest release: 2.2.2_0.11.0 (2018-12-23) /Apache-2.0 / (__10)
* 3testing
* 1streaming
* 1tools
*
-------------------------PYSPARK-CSV
AN EXTERNAL PYSPARK MODULE THAT WORKS LIKE R'S READ.CSV OR PANDA'S READ_CSV, WITH AUTOMATIC TYPE INFERENCE AND NULL VALUE HANDLING. PARSES CSV DATA INTO SCHEMARDD. NO INSTALLATION REQUIRED, SIMPLY INCLUDE PYSPARK_CSV.PY VIA SPARKCONTEXT. @seahboonsiew / No release yet / (__1)* 2python
* 2csv
* 1sql
*
-------------------------SPARK-MONGODB
MONGODB DATA SOURCE FOR SPARK SQL @Stratio / Latest release: 0.12.0 (2016-08-31) /Apache-2.0 / (__14)
* 5MongoDB
* 5Spark SQL
* 2sql
*
-------------------------PYSPARK-CASSANDRA
PYSPARK CASSANDRA BRINGS BACK THE FUN IN WORKING WITH CASSANDRA DATAIN PYSPARK.
@TargetHolding / Latest release: 0.3.5 (2016-03-30) / Apache-2.0 / (__1)* 1python
* 1spark
* 1sql
*
-------------------------SPARK_DBSCAN
DBSCAN CLUSTERING ALGORITHM ON TOP OF APACHE SPARK @alitouka / No release yet / (__0)* 1clustering
* 1ml
* 1dbscan
*
-------------------------DEMO-SCALA-PYTHON
A SPARK PACKAGE TEMPLATE @brkyvz / Latest release: 1.2-s_2.10 (2016-05-25) /Apache-2.0 / (__1)
* 1python
* 1demo
* 1template
*
-------------------------SBT-SPARK-PACKAGE
SBT PLUGIN FOR SPARK PACKAGES @databricks / Latest release: 0.2.4 (2016-07-15) /Apache-2.0 / (__3)
* 1tools
* 1sbt
*
-------------------------SPARK-SCALDING
USE CASCADING TAPS AND SCALDING DSL WITH SPARK — EDIT @tresata / Latest release: 0.5.0-s_2.10 (2015-11-13) /Apache-2.0 / (__0)
-------------------------SPARK-SORTED
SECONDARY SORT AND STREAMING REDUCE FOR SPARK @tresata / Latest release: 0.4.0-s_2.11 (2015-11-03) /Apache-2.0 / (__0)
* 1core
*
-------------------------SPARK-KAFKA
LOW LEVEL INTEGRATION OF SPARK AND KAFKA @tresata / Latest release: 0.6.0-s_2.10 (2015-11-13) /Apache-2.0 / (__0)
* 1streaming
*
------------------------- SPARK-CASSANDRA-CONNECTOR CONNECTS SPARK TO CASSANDRA @datastax / Latest release: 2.4.0-s_2.11 (2018-11-29) / Apache-2.0 / (__14)* 3spark
* 3cassandra
* 2nosql
*
------------------------- SPARK-PACKAGE-CMD-TOOL A COMMAND LINE TOOL FOR SPARK PACKAGES @databricks / Latest release: 0.3.0 (2015-03-17) /Apache-2.0 / (__1)
* 1tools
*
-------------------------SPAWNCAMPING-DDS
DATA-DRIVEN SPARK ALLOWS QUICK DATA EXPLORATION BASED ON APACHE SPARK @FRosner / No release yet / (__0) -------------------------SPARK-POWER-BI
POWER BI API ADAPTER FOR APACHE SPARK @granturing / Latest release: 1.5.0_0.0.7 (2015-09-13) / Apache-2.0 / (__0)* 2streaming
* 1sql
* 1realtime
*
-------------------------MEETUP-STREAM
SPARK STREAMING, MACHINE LEARNING AND MEETUP.COM STREAMING API. @actions / No release yet / (__1)* 1ml
* 1example
* 1streaming
*
-------------------------DEMO-PYTHON-SP
PURE PYTHON PACKAGE USED FOR TESTING SPARK PACKAGES @brkyvz / Latest release: 0.4.2 (2016-02-14) /Apache-2.0 / (__0)
------------------------- SPARK-MRMR-FEATURE-SELECTION FEATURE SELECTION BASED ON INFORMATION GAIN: MAXIMUM RELEVANCY MINIMUMREDUNDANCY
@wxhC3SC6OPm8M1HXboMy / No release yet /(__0)
* 1mllib
*
-------------------------SPARK-CSV2SQL
HAND ROUTINE TO IMPORT CSV FILES AS TABLES IN SPARK SQL @wxhC3SC6OPm8M1HXboMy / No release yet /(__0)
-------------------------SPARK-SEQUOIADB
SPARK CONNECTOR FOR SEQUOIADB @SequoiaDB / Latest release: 1.12-s_2.11 (2015-03-30) / Apache-2.0 / (__2)* 2sequoiadb
* 2nosql
* 2sql
*
-------------------------SPARK-NOTEBOOK
USE APACHE SPARK STRAIGHT FROM THE BROWSER @andypetrella / Latest release: v0.4.0 (2015-03-29) / Apache-2.0 / (__2)* 1notebook
* 1charts
* 1interactive
*
-------------------------Next >
------------------------- Spark Packages is a community site hosting modules that are not part of Apache Spark. Your use of and access to this site is subject to theterms of use .
Apache Spark and the Spark logo are trademarks of the Apache Software Foundation. This site is maintained as a community service byDatabricks .
Details
Copyright © 2024 ArchiveBay.com. All rights reserved. Terms of Use | Privacy Policy | DMCA | 2021 | Feedback | Advertising | RSS 2.0