Skip to content

Commit

Permalink
Update API links. Fix Hive example.
Browse files Browse the repository at this point in the history
  • Loading branch information
marmbrus authored and ahirreddy committed Apr 15, 2014
1 parent 58e2aa9 commit 227a0be
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions docs/sql-programming-guide.md
Original file line number Diff line number Diff line change
Expand Up @@ -38,8 +38,8 @@ file, or by running HiveQL against data stored in [Apache Hive](http://hive.apac

Spark SQL allows relational queries expressed in SQL or HiveQL to be executed using
Spark. At the core of this component is a new type of RDD,
[SchemaRDD](). SchemaRDDs are composed
[Row]() objects along with
[SchemaRDD](api/pyspark/pyspark.rdd.SchemaRDD-class.html). SchemaRDDs are composed
[Row](api/pyspark/pyspark.rdd.Row-class.html) objects along with
a schema that describes the data types of each column in the row. A SchemaRDD is similar to a table
in a traditional relational database. A SchemaRDD can be created from an existing RDD, parquet
file, or by running HiveQL against data stored in [Apache Hive](http://hive.apache.org/).
Expand Down Expand Up @@ -85,7 +85,7 @@ JavaSQLContext sqlCtx = new org.apache.spark.sql.api.java.JavaSQLContext(ctx);
<div data-lang="python" markdown="1">

The entry point into all relational functionality in Spark is the
[SQLContext]() class, or one
[SQLContext](api/pyspark/pyspark.context.SQLContext-class.html) class, or one
of its decedents. To create a basic SQLContext, all you need is a SparkContext.

{% highlight python %}
Expand Down Expand Up @@ -401,7 +401,7 @@ expressed in HiveQL.
{% highlight python %}

from pyspark.context import HiveContext
hiveCtx = HiveContext(sqlCtx)
hiveCtx = HiveContext(sc)

hiveCtx.hql("CREATE TABLE IF NOT EXISTS src (key INT, value STRING)")
hiveCtx.hql("LOAD DATA LOCAL INPATH 'examples/src/main/resources/kv1.txt' INTO TABLE src")
Expand Down

0 comments on commit 227a0be

Please sign in to comment.