R interface for Apache Spark
-
Updated
Nov 19, 2025 - R
Apache Spark is an open source distributed general-purpose cluster-computing framework. It provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
R interface for Apache Spark
bring sf to spark in production
R interface for XGBoost on Spark
R interface to Spark TensorFlow Connector
Sparklyr extension making Flint time series library functionalities (https://github.com/twosigma/flint) easily accessible through R
Old repo for R interface for GraphFrames
A sparklyr extension to analyze genome datasets
Enable spatial functions in Spark through the `sparklyr` package
This repository you are browsing contains intermediate level piece of codes which are useful for cleaning, exploratory analysis, handling of missing data points, outlier detection and different visualization techniques using graphics, ggplot2, tidycharts, ggExtra packages. Also in particular part of the script you can get basic information about…
R workloads running at scale on Google Cloud
A tutorial showing how to use Apache Spark, Apache Sedona, and Delta Lake for big data analysis in R.
Using Apache Spark for marketing analytics
Mirror of https://gitlab.com/zero323/dlt
Projects created using R
Created by Matei Zaharia
Released May 26, 2014