我正在尝试使用 Zeppelin(Windows 10 上的 v 0.7.0 java 1.8;与 docker<相同)/a> v .0.7.1) 用于查询 Oracle 数据库的 JDBC 解释器.
I am trying to use Zeppelin (v 0.7.0 java 1.8 on Windows 10; same with docker v .0.7.1) JDBC interpreter to query Oracle Database.
到目前为止,我已经找到了诸如 example 之类的论文.我尝试使用 jdbc 解释器:
So far I've found papers like example. I try to use jdbc interpreter with:
common.max_count=100
default.driver=oracle.jdbc.pool.OracleDataSource
default.password:$password
default.user=$my_user_name
default.url=jdbc:oracle:thin:@$host:1521/$service_name>
zeppelin.jdbc.concurrent.max_connection=10
zeppelin.jdbc.concurrent.use=true
连接看起来已成功建立*,但即使是最简单的查询(如sql% select 1 from dual)也会永远运行.
Connection looks to be established successfully*, but even simplest query (like sql% select 1 from dual) runs for ever.
有人知道/知道要添加什么吗?
Does anyone knows/have a clue on what is to be added?
我在终端看到的 - 重复行
what I see in terminal - repeating lines
sKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbcDEBUG [2017-06-15 16:03:28,203] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbcDEBUG [2017-06-15 16:03:28,204] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbcDEBUG [2017-06-15 16:03:28,204] ({Thread-90} InterpreterFactory.java[getInterpreterSessionKey]:1244) - 解释器会话密钥:shared_session,注意:2CMEC9SS4,用户:匿名,解释器设置名称:jdbcDEBUG [2017-06-15 16:03:28,205] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbcDEBUG [2017-06-15 16:03:28,205] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc
sKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc DEBUG [2017-06-15 16:03:28,203] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc DEBUG [2017-06-15 16:03:28,204] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc DEBUG [2017-06-15 16:03:28,204] ({Thread-90} InterpreterFactory.java[getInterpreterSessionKey]:1244) - Interpreter session key: shared_session, for note: 2CMEC9SS4, user: anonymous, InterpreterSetting Name: jdbc DEBUG [2017-06-15 16:03:28,205] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc DEBUG [2017-06-15 16:03:28,205] ({Thread-90} InterpreterSetting.java[getInterpreterProcessKey]:143) - getInterpreterProcessKey: shared_process for InterpreterSetting Id: 2CMBZV4T1, Name: jdbc
与此同时,我在使用带有 Scala 的 spark 解释器而不是 jdbc 相同的驱动程序时没有问题.
In the same time I don't have an issue using spark interpretor with scala over jdbc same driver.
您确定查询正在运行吗?
Are you sure a query is running?
selec 1 from dual
不是有效的 SQL.
在 gv$session 中检查 Oracle 端的查询.如果它正在运行,它就会在那里.
Check on Oracle side in gv$session for your query. If it is running, it will be there.
这篇关于Zeppelin Oracle SQL 查询优先运行的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持html5模板网!
Pyspark DataFrameWriter jdbc 函数的 ignore 选项是忽略整Does ignore option of Pyspark DataFrameWriter jdbc function ignore entire transaction or just offending rows?(Pyspark DataFrameWriter jdbc 函数的 ig
pyspark mysql jdbc load 调用 o23.load 时发生错误 没有合pyspark mysql jdbc load An error occurred while calling o23.load No suitable driver(pyspark mysql jdbc load 调用 o23.load 时发生错误 没有合适的
在 Apache Spark 2.0.0 中,是否可以从外部数据库获取In Apache Spark 2.0.0, is it possible to fetch a query from an external database (rather than grab the whole table)?(在 Apache Spark 2.0.0 中,是否可
Spark在执行jdbc保存时给出空指针异常Spark giving Null Pointer Exception while performing jdbc save(Spark在执行jdbc保存时给出空指针异常)
Pyspark DataFrameWriter jdbc 函数的 ignore 选项是忽略整Does ignore option of Pyspark DataFrameWriter jdbc function ignore entire transaction or just offending rows?(Pyspark DataFrameWriter jdbc 函数的 ig
在 Spark 中找不到适合 jdbc 的驱动程序No suitable driver found for jdbc in Spark(在 Spark 中找不到适合 jdbc 的驱动程序)