日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

hive整合phoenix

發布時間:2025/3/21 编程问答 22 豆豆
生活随笔 收集整理的這篇文章主要介紹了 hive整合phoenix 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

2019獨角獸企業重金招聘Python工程師標準>>>

版本:

hbase-0.98.21-hadoop2-bin.tar.gz

phoenix-4.8.0-HBase-0.98-bin.tar.gz

apache-hive-1.2.1-bin.tar.gz

--------------------------------------------------

首先需要phoenix整合hbase

hive整合hbase,此處參照之前的筆記

將phoenix{core,queryserver,4.8.0-HBase-0.98,hive}拷貝到$hive/lib/

根據官網要求修改配置文件

> vim conf/hive-env.sh

> vim conf/hive-site.xml

啟動:

> hive -hiveconf phoenix.zookeeper.quorum=hadoop01:2181

創建內部表

create table phoenix_table (

s1 string,

i1 int,

f1 float,

d1 double

)

STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler'

TBLPROPERTIES (

"phoenix.table.name" = "phoenix_table",

"phoenix.zookeeper.quorum" = "hadoop01",

"phoenix.zookeeper.znode.parent" = "/hbase",

"phoenix.zookeeper.client.port" = "2181",

"phoenix.rowkeys" = "s1, i1",

"phoenix.column.mapping" = "s1:s1, i1:i1, f1:f1, d1:d1",

"phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'"

);

創建成功。查詢phoenix和hbase中都有相應的表生成:phoenix

hbase:

屬性

  • phoenix.table.name
      • phoenix指定表名
      • 默認值:hive一樣的表
  • phoenix.zookeeper.quorum
      • 指定ZK地址
      • 默認值:localhost
  • phoenix.zookeeper.znode.parent
      • 指定HBase在ZK的目錄
      • 默認值:/ hbase
  • phoenix.zookeeper.client.port
      • 指定ZK端口
      • 默認值:2181
  • phoenix.rowkeys
      • 指定phoenix的rowkey,即hbase的rowkey
      • 要求
  • phoenix.column.mapping
      • hive與phoenix之間的列映射。

    插入數據

    使用hive測試表pokes導入數據

    > insert into table phoenix_table select bar,foo,12.3 as fl,22.2 as dl from pokes;

    成功、查詢

    在phoenix中查詢

    還可以使用phoenix導入數據,看官網的解釋

    注意:phoenix4.8認為加tbale關鍵字為語法錯誤,其他版本沒試,不知道官網怎么沒說明

    創建外部表

    For external tables Hive works with an existing Phoenix table and manages only Hive metadata. Deleting an external table from Hive only deletes Hive metadata and keeps Phoenix table

    首先在phoenix創建表

    phoenix> create table PHOENIX_TABLE_EXT(aa varchar not null primary key,bb varchar);

    再在hive中創建外部表:

    create external table phoenix_table_ext_1 ( aa string, bb string ) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ( "phoenix.table.name" = "phoenix_table_ext ", "phoenix.zookeeper.quorum" = "hadoop01", "phoenix.zookeeper.znode.parent" = "/hbase", "phoenix.zookeeper.client.port" = "2181", "phoenix.rowkeys" = "aa", "phoenix.column.mapping" = "aa:aa, bb:bb" );

    創建成功,插入成功

    這些選項可以設置在hive CLI

    性能調優

    參數默認值描述
    phoenix.upsert.batch.size1000批量大小插入。
    [phoenix-table-name].disable.walfalse它暫時設置表屬性DISABLE_WAL = true??捎糜谔岣咝阅?/td>
    [phoenix-table-name].auto.flushfalse當WAL是disabled 的flush又為真,則按文件刷進庫

    查詢數據

    可以使用HiveQL在phoenix表查詢數據。一個簡單表查詢當hive.fetch.task.conversion=more?and?hive.exec.parallel=true.就可以像在Phoenix CLI一樣快。

    參數默認值描述
    hbase.scan.cache100為一個單位請求讀取行大小。
    hbase.scan.cacheblockfalse是否緩存塊。
    split.by.statsfalseIf true, mappers will use table statistics. One mapper per guide post.
    [hive-table-name].reducer.count1reducer的數量. In tez mode is affected only single-table query. See Limitations
    [phoenix-table-name].query.hint?Hint for phoenix query (like NO_INDEX)

    遇到的問題:

    FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hbase.client.Scan.isReversed()Z

    最開始我用的hbase-0.96.2-hadoop2版本,不能整合,這個是需要hbase-client-0.98.21-hadoop2.jar包,更換這個jar包就解決了,但是還是會報下面的錯

    FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:ERROR 103 (08004): Unable to establish connection.

    于是更換了hbase的版本為0.98.21的 ok了

    ---------

    FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. java.lang.StringIndexOutOfBoundsException: String index out of range: -1

    因為字段對應不一樣

    create table phoenix_table_3 (a string,b int) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ("phoenix.table.name" = "phoenix_table_3","phoenix.zookeeper.quorum" = "hadoop01","phoenix.zookeeper.znode.parent" = "/hbase","phoenix.zookeeper.client.port" = "2181","phoenix.rowkeys" = "a1","phoenix.column.mapping" = "a:a1, b:b1","phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'");

    hive表字段與phoenix字段一樣就可以了

    ----------

    創建成功,插入也能成功,就是hive查詢的時候報錯找不到a1列,因為phoenix是aa列

    Failed with exception java.io.IOException:java.lang.RuntimeException: org.apache.phoenix.schema.ColumnNotFoundException: ERROR 504 (42703): Undefined column. columnName=A1

    create external table phoenix_table_ext (a1 string,b1 string)STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ("phoenix.table.name" = "phoenix_table_ext","phoenix.zookeeper.quorum" = "hadoop01","phoenix.zookeeper.znode.parent" = "/hbase","phoenix.zookeeper.client.port" = "2181","phoenix.rowkeys" = "aa","phoenix.column.mapping" = "a1:aa, b1:bb");

    解決辦法:同上hive表字段與phoenix字段一樣就可以了

    轉載于:https://my.oschina.net/zlhblogs/blog/783894

    總結

    以上是生活随笔為你收集整理的hive整合phoenix的全部內容,希望文章能夠幫你解決所遇到的問題。

    如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。