top of page

Write CSV/JSON data to Elasticsearch using Spark dataframes

Updated: Nov 18, 2022

Elasticsearch-hadoop connector allows Spark-elasticsearch integration in Scala and Java language. Elasticsearch-hadoop library helps Apache Spark to integrate with Elasticsearch.


  • Write JSON data to Elasticsearch using Spark dataframe

  • Write CSV file to Elasticsearch using Spark dataframe

I am using Elasticsearch version [7.3.0], Spark [2.3.1] and Scala [2.11].


Download Jar

In order to execute Spark with Elasticsearch, you need to download proper version of spark-elasticsearch jar file and add it to Spark's classpath. If you are running Spark in local mode it will be added to just one machine but if you are running in cluster, you need to add it per-node.

I assume you have already installed Elasticsearch, if not please follow these for installation steps (Linux | Mac users). Elasticsearch installation is very easy and it will be done in few minutes. I would encourage you all to install Kibana as well.

Now, you can download complete list of hadoop library (Storm, Mapreduce, Hive and Pig as shown below) from here. I have added elasticsearch-spark-20_2.10-7.3.0.jar because I am running Elastics 7.3 version.

[Tip] Make sure you are downloading correct version of jar, otherwise you will get this error during execution: org.elasticsearch.hadoop.EsHadoopIllegalArgumentException: Unsupported/Unknown Elasticsearch version x.x.x


Adding Jar (Scala IDE)

If you are using Scala IDE, just right click on project folder => go to properties => Java build path => add external jars and add the downloaded jar file. Apply and close.


Adding Jar (Spark-shell)

If you are using Spark-shell, just navigate to the Spark executable library where you can see all other jar files and add the downloaded jar file there. For example,


Start Elasticsearch & Kibana

Now, make sure Elasticsearch is running. If Elasticsearch is not running, Spark will not be able to make connection and you will get this error. Connection error (check network and/or proxy settings)- all nodes failed.

To start Elasticsearch and Kibana run this command on your terminal,

$ elasticsearch

$ kibana


Writing JSON data to Elasticsearch

In all sections these three steps are mandatory,

  • Import necessary elasticsearch spark library

  • Configure ES nodes

  • Configure ES port

  • If you are running ES on AWS just add this line to your configurations - .config("","true")

JSON file

multilinecolors.json sample data:

[ { "color": "red", "value": "#f00" }, { "color": "green", "value": "#0f0" }, { "color": "blue", "value": "#00f" }, { "color": "cyan", "value": "#0ff" }, { "color": "magenta", "value": "#f0f" }, { "color": "yellow", "value": "#ff0" }, { "color": "black", "value": "#000" } ]

package com.dataneb.spark

import org.apache.spark.sql.SparkSession

import org.elasticsearch.spark.sql._

object toES {

def main(args: Array[String]): Unit = {

// Configuration

val spark = SparkSession







// Create dataframe

val colorsDF ="/Volumes/MYLAB/testdata/multilinecolors.json")

// Write to ES with index name in lower case