Hive MetaStore Upgrade

# cd $HIVE_HOME/scripts/metastore/upgrade/mysql
[Dev root @ sd-9c1f-2eac /usr/local/src/apache-hive-2.1.1-bin/scripts/metastore/upgrade/mysql]
# ls
001-HIVE-972.mysql.sql 027-HIVE-12819.mysql.sql hive-schema-2.0.0.mysql.sql
002-HIVE-1068.mysql.sql 028-HIVE-12821.mysql.sql hive-schema-2.1.0.mysql.sql
003-HIVE-675.mysql.sql 029-HIVE-12822.mysql.sql hive-txn-schema-0.13.0.mysql.sql
004-HIVE-1364.mysql.sql 030-HIVE-12823.mysql.sql hive-txn-schema-0.14.0.mysql.sql
005-HIVE-417.mysql.sql 031-HIVE-12831.mysql.sql hive-txn-schema-1.3.0.mysql.sql
006-HIVE-1823.mysql.sql 032-HIVE-12832.mysql.sql hive-txn-schema-2.0.0.mysql.sql
007-HIVE-78.mysql.sql 034-HIVE-13076.mysql.sql hive-txn-schema-2.1.0.mysql.sql
008-HIVE-2246.mysql.sql 035-HIVE-13395.mysql.sql README
009-HIVE-2215.mysql.sql 036-HIVE-13354.mysql.sql upgrade-0.10.0-to-0.11.0.mysql.sql
010-HIVE-3072.mysql.sql hive-schema-0.10.0.mysql.sql upgrade-0.11.0-to-0.12.0.mysql.sql
011-HIVE-3649.mysql.sql hive-schema-0.11.0.mysql.sql upgrade-0.12.0-to-0.13.0.mysql.sql
012-HIVE-1362.mysql.sql hive-schema-0.12.0.mysql.sql upgrade-0.13.0-to-0.14.0.mysql.sql
013-HIVE-3255.mysql.sql hive-schema-0.13.0.mysql.sql upgrade-0.14.0-to-1.1.0.mysql.sql
014-HIVE-3764.mysql.sql hive-schema-0.14.0.mysql.sql upgrade-0.5.0-to-0.6.0.mysql.sql
016-HIVE-6386.mysql.sql hive-schema-0.3.0.mysql.sql upgrade-0.6.0-to-0.7.0.mysql.sql
017-HIVE-6458.mysql.sql hive-schema-0.4.0.mysql.sql upgrade-0.7.0-to-0.8.0.mysql.sql
018-HIVE-6757.mysql.sql hive-schema-0.4.1.mysql.sql upgrade-0.8.0-to-0.9.0.mysql.sql
019-HIVE-7784.mysql.sql hive-schema-0.5.0.mysql.sql upgrade-0.9.0-to-0.10.0.mysql.sql
020-HIVE-9296.mysql.sql hive-schema-0.6.0.mysql.sql upgrade-1.1.0-to-1.2.0.mysql.sql
021-HIVE-7018.mysql.sql hive-schema-0.7.0.mysql.sql upgrade-1.2.0-to-1.3.0.mysql.sql
022-HIVE-11970.mysql.sql hive-schema-0.8.0.mysql.sql upgrade-1.2.0-to-2.0.0.mysql.sql
023-HIVE-12807.mysql.sql hive-schema-0.9.0.mysql.sql upgrade-2.0.0-to-2.1.0.mysql.sql
024-HIVE-12814.mysql.sql hive-schema-1.1.0.mysql.sql upgrade.order.mysql
025-HIVE-12816.mysql.sql hive-schema-1.2.0.mysql.sql
026-HIVE-12818.mysql.sql hive-schema-1.3.0.mysql.sql

cat README
Hive MetaStore Upgrade HowTo
============================

This document describes how to upgrade the schema of a MySQL backed
Hive MetaStore instance from one release version of Hive to another
release version of Hive. For example, by following the steps listed
below it is possible to upgrade a Hive 0.5.0 MetaStore schema to a
Hive 0.7.0 MetaStore schema. Before attempting this project we
strongly recommend that you read through all of the steps in this
document and familiarize yourself with the required tools.

MetaStore Upgrade Steps
=======================

1) Shutdown your MetaStore instance and restrict access to the
MetaStore‘s MySQL database. It is very important that no one else
accesses or modifies the contents of database while you are
performing the schema upgrade.

2) Create a backup of your MySQL metastore database. This will allow
you to revert any changes made during the upgrade process if
something goes wrong. The mysqldump utility is the easiest way to
create a backup of a MySQL database:

% mysqldump --opt <metastore_db_name> > metastore_backup.sql

Note that you may need also need to specify a hostname and username
using the --host and --user command line switches.

3) Dump your metastore database schema to a file. We use the mysqldump
utility again, but this time with a command line option that
specifies we are only interested in dumping the DDL statements
required to create the schema:

% mysqldump --skip-add-drop-table --no-data <metastore_db_name> > my-schema-x.y.z.mysql.sql

4) The schema upgrade scripts assume that the schema you are upgrading
closely matches the official schema for your particular version of
Hive. The files in this directory with names like
"hive-schema-x.y.z.mysql.sql" contain dumps of the official schemas
corresponding to each of the released versions of Hive. You can
determine differences between your schema and the official schema
by diffing the contents of the official dump with the schema dump
you created in the previous step. Some differences are acceptable
and will not interfere with the upgrade process, but others need to
be resolved manually or the upgrade scripts will fail to complete.

* Missing Tables: Hive‘s default configuration causes the MetaStore
to create schema elements only when they are needed. Some tables
may be missing from your MetaStore schema if you have not created
the corresponding Hive catalog objects, e.g. the PARTITIONS table
will probably not exist if you have not created any table
partitions in your MetaStore. You MUST create these missing tables
before running the upgrade scripts. The easiest way to do this is
by executing the official schema DDL script against your
schema. Each of the CREATE TABLE statements in the schema script
include an IF NOT EXISTS clause, so tables which already exist in
your schema will be ignored, and those which don‘t exist will get
created.

* Extra Tables: Your schema may include a table named NUCLEUS_TABLES
or a table named SEQUENCE_TABLE. These tables are managed
by the DataNucleus ORM layer and will be created automatically if
they don‘t exist. No action on your part is required.

* Reversed Column Constraint Names in the Same Table: Tables with
multiple constraints may have the names of the constraints
reversed. For example, the PARTITIONS table contains two foreign
key constraints named PARTITIONS_FK1 and PARTITIONS_FK2 which
reference SDS.SD_ID and TBLS.TBL_ID respectively. However, in your
schema you may find that PARTITIONS_FK1 references TBLS.TBL_ID and
PARTITIONS_FK2 references SDS.SD_ID. Either version is acceptable
-- the only requirement is that these constraints actually exist.

* Differences in Column/Constraint Names: Your schema may contain
tables with columns named "IDX" or unique keys named
"UNIQUE<tab_name>". If you find either of these in your schema you
will need to change the names to "INTEGER_IDX" and
"UNIQUE_<tab_name>" before running the upgrade scripts. For more
background on this issue please refer to HIVE-1435.

5) You are now ready to run the schema upgrade scripts. If you are
upgrading from Hive 0.5.0 to Hive 0.6.0 you need to run the
upgrade-0.5.0-to-0.6.0.mysql.sql script, but if you are upgrading
from 0.5.0 to 0.7.0 you will need to run the 0.5.0 to 0.6.0 upgrade
script followed by the 0.6.0 to 0.7.0 upgrade script.

% mysql --verbose
mysql> use <metastore_db_name>;
Database changed
mysql> source upgrade-0.5.0-to-0.6.0.mysql.sql
mysql> source upgrade-0.6.0-to-0.7.0.mysql.sql

These scripts should run to completion without any errors. If you
do encounter errors you need to analyze the cause and attempt to
trace it back to one of the preceding steps.

6) The final step of the upgrade process is validating your freshly
upgraded schema against the official schema for your particular
version of Hive. This is accomplished by repeating steps (3) and
(4), but this time comparing against the official version of the
upgraded schema, e.g. if you upgraded the schema to Hive 0.7.0 then
you will want to compare your schema dump against the contents of
hive-schema-0.7.0.mysql.sql

时间: 2024-10-12 19:06:18

Hive MetaStore Upgrade的相关文章

【原创】大叔经验分享(23)hive metastore的几种部署方式

hive及其他组件(比如spark.impala等)都会依赖hive metastore,依赖的配置文件位于hive-site.xml hive metastore重要配置 hive.metastore.warehouse.dirhive2及之前版本默认为/user/hive/warehouse/,创建数据库或表时会在该目录下创建对应的目录 javax.jdo.option.ConnectionURLjavax.jdo.option.ConnectionDriverNamejavax.jdo.o

Hive Metastore 连接报错

背景 项目中需要通过一些自定义的组件来操控hive的元数据,于是使用了remote方式来存储hive元数据,使用一个服务后台作为gateway,由它来控制hive元数据. 现象 在windows上连接hive metastore的时候,无端的会报NullPointerException,非常费解. 分析 看了代码后发现,连接后会获取本地用户所在的用户组信息(org.apache.hadoop.hive.metastore.HiveMetaStoreClient中的open方法): if (isC

hive metastore Server 出现异常

报错信息: 常见问题分析: 1 hive metastore 数据库中用户名或者密码出现更改,并且重启了hive,导致生效但是CDH下没有及时更改hive metastore设置密码 2 Mysql 的连接驱动" mysql-connector-java-5.1.26-bin.jar "缺失.该驱动 需要同时存在于  $JAVA_HOME/lib/目录下.$HIVE_HOME/lib/目录下. 并且设置755权限

hive 报错 java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient

Exception in thread "main" java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:344) at org.a

Hive metastore源码阅读(一)

不要问我为什么,因为爱,哈哈哈哈...进入正题,最近做项目顺带学习了下hive metastore的源码,进行下知识总结. hive metastore的整体架构如图: 一.组成结构: 如图我们可以看到,hive metastore的组成结构分为 客户端 服务端 ,那么下来我们逐一进行分析: 1.客户端  从代码的角度来看:尼玛太多了..我们从入口HIVE开始看,可以找到MetaStoreClient客户端的创建: 1 private IMetaStoreClient createMetaSto

Hive metastore三种配置方式

本文转载至:http://blog.csdn.net/reesun/article/details/8556078 目录(?)[-] 一本地derby 二本地mysql 三远端mysql 1服务端配置文件 2客户端配置文件 Hive的meta数据支持以下三种存储方式,其中两种属于本地存储,一种为远端存储.远端存储比较适合生产环境.Hive官方wiki详细介绍了这三种方式,链接为:Hive Metastore. 一.本地derby 这种方式是最简单的存储方式,只需要在hive-site.xml做如

hive metastore异常 org.apache.thrift.protocol.TProtocolException: Missing version in readMessageBegin, old client

hiveserver2的端口是10000hive.metastoe.uris 的端口9083改为10000之后 beelien 连接hiveserver2报错 Error: Could not open client transport with JDBC Uri: jdbc:hive2://172.16.145.124:10000 hive 用户和密码都没数错,试了好几次还是不能连接 查看hive  metastore的日志,发现有错误信息: 想了想是不是因为metastore和hiveser

shematool -initschema -dbtype mysql error org.apache.hadoop.hive.metastore.hivemetaexception:Failed to get schema version

命令:schematool -initSchema -dbType mysql Fix the issue: edit /etc/mysql/my.cnf change bind-address            = 127.0.0.1    to  bind-address            = 0.0.0.0 Restart mysqld (/etc/init.d/mysql restart).

Hive的metastore

hive --service metastore 默认端口是9083 <property> <name>hive.metastore.uris</name> <value>thrift://hiveserver1:9083</value> </property> 在连接hive的客户端,如spark,可以只在hive-site.xml中配置这一个属性即可. 不需要配置mysql连接信息与用户密码信息,hiver server作为thr