Spark Programming Spark SQL

Spark Programming ? Spark SQL

Bu eitim sunumlari stanbul Kalkinma Ajansi'nin 2016 yili Yenilik?i ve Yaratici stanbul Mali Destek Programi kapsaminda y?r?t?lmekte olan TR10/16/YNY/0036 no'lu stanbul Big Data Eitim ve Aratirma Merkezi Projesi dahilinde ger?ekletirilmitir. ?erik ile ilgili tek sorumluluk Bah?eehir ?niversitesi'ne ait olup STKA veya Kalkinma Bakanlii'nin g?r?lerini yansitmamaktadir.

Spark SQL

blurs the lines between RDDs and relational tables

intermix SQL commands to query external data, along with complex analytics, in a single app:

? allows SQL extensions based on MLlib ? Shark is being migrated to Spark SQL

Spark SQL

val sqlContext = new org.apache.spark.sql.SQLContext(sc) import sqlContext._

// Define the schema using a case class. case class Person(name: String, age: Int)

// Create an RDD of Person objects and register it as a table. val people = sc.textFile("examples/src/main/resources/ people.txt").map(_.split(",")).map(p => Person(p(0), p(1).trim.toInt))

people.registerAsTable("people")

// SQL statements can be run by using the sql methods provided by sqlContext. val teenagers = sql("SELECT name FROM people WHERE age >= 13 AND age ................
................

In order to avoid copyright disputes, this page is only a partial summary.

Google Online Preview   Download