Explain pipe() operation in Apache Spark
-
-
Explain pipe() operation in Apache Spark
-
<li style=”list-style-type: none”>
def pipe(command: String): RDD[String]
Return an RDD created by piping elements to a forked external process.
- In general, Spark is using Scala, Java, and Python to write the program. However, if that is not enough, and one want to pipe (inject) the data which written in other languages like ‘R’, Spark provides general mechanism in the form of pipe() method
- Spark provides the pipe() method on RDDs.
- With Spark’s pipe() method, one can write a transformation of an RDD that can read each element in the RDD from standard input as String.
- It can write the results as String to the standard output.
For more transformation on RDDs see: Apache Spark Operations
- You must be logged in to reply to this topic.