From 665852d6b6e8a011a32055f16aaf40a5c3a9d45f Mon Sep 17 00:00:00 2001 From: Jaume Marhuenda Date: Mon, 9 Nov 2020 06:55:36 -0500 Subject: [PATCH] update readme with spark (#127) --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index cedb885..c38c172 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,7 @@ [![Build Status](https://travis-ci.org/beltran/gohive.svg?branch=master)](https://travis-ci.org/beltran/gohive) [![Coverage Status](https://coveralls.io/repos/github/beltran/gohive/badge.svg?branch=master)](https://coveralls.io/github/beltran/gohive?branch=master) -GoHive is a driver for Hive in go that supports connection mechanisms KERBEROS(Gssapi Sasl), NONE(Plain Sasl), LDAP, CUSTOM and NOSASL, both for binary and http transport, with and without SSL. The kerberos mechanism will pick a different authentication level depending on `hive.server2.thrift.sasl.qop`. +GoHive is a driver for Hive and the [Spark Distributed SQL Engine](https://spark.apache.org/docs/latest/sql-distributed-sql-engine.html) in go that supports connection mechanisms KERBEROS(Gssapi Sasl), NONE(Plain Sasl), LDAP, CUSTOM and NOSASL, both for binary and http transport, with and without SSL. The kerberos mechanism will pick a different authentication level depending on `hive.server2.thrift.sasl.qop`. ## Installation Gohive can be installed with: @@ -128,7 +128,7 @@ cursor.FetchOne(context.Background(), &i) ``` Alternatively, using the rowmap API, `m := cursor.RowMap(context.Background())`, `m` would be `map[string]interface{}{"table_name.column_name": nil}` for a `NULL` value. It will return a map -where the keys are `table_name.column_name`. This works fine with hive but using [Spark Thirft SQL server](https://github.com/apache/spark/blob/master/sbin/start-thriftserver.sh) `table_name` is not present and the keys are `column_name` and it can [lead to problems](https://github.com/beltran/gohive/issues/120) if two tables have the same column name so the `FetchOne` API should be used in this case. +where the keys are `table_name.column_name`. This works fine with hive but using [Spark Thirft SQL server](https://spark.apache.org/docs/latest/sql-distributed-sql-engine.html) `table_name` is not present and the keys are `column_name` and it can [lead to problems](https://github.com/beltran/gohive/issues/120) if two tables have the same column name so the `FetchOne` API should be used in this case. ## Running tests Tests can be run with: