Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient

Caused by: org.datanucleus.exceptions.NucleusException: Attempt to invoke the "DBCP" plugin to create a ConnectionPool gave an error : The specified datastore driver ("com.mysql.jdbc.Driver") was not found in the CLASSPATH. Please check your CLASSPATH specification, and the name of the driver.


解决方案一:查了好多都没有解决由于stackoverflow回复出了点bug只好在自己的博客记录了。

I meet the same problems !

All of the config goes well ,but the mysql-connector-java-5.1.17.jar seems to not work.

Now try this way ,specify the jars when you run spark shell.



spark-shell --jars /usr/local/hive-0.13.1/lib/mysql-connector-java-5.1.17.jar 

En ,you can try hive shell as this way.


直接:添加 --jars   指定mysql-connector-java-5.1.17.jar 



方案二: 

将该jar包上传到hdfs,然后 --jars指定


方案三:

将该jar包上传到各个节点上,并在 /etc/alternatives/spark-conf/classpath.txt 指定

/etc/alternatives/spark-conf/classpath.txt