使用 Druid、Kafka、Superset 进行流式传输
Streaming with Druid, Kafka, Superset
我正在使用 Kafka、Druid 和 SuperSet 测试数据流。
我在 Druid 中有一些数据(参见 1.picutre)。
之后我可以通过选项 "Refresh Druid metadata" 在 Superset 中生成 Druid 数据源(参见 2.pic)
问题是当我想查询数据时,我收到此错误消息:
URLError: <urlopen error [Errno -2] Name or service not known>
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/superset/viz.py", line 329, in get_df_payload
df = self.get_df(query_obj)
File "/usr/lib/python2.7/site-packages/superset/viz.py", line 142, in get_df
self.results = self.datasource.query(query_obj)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 1238, in query
client=client, query_obj=query_obj, phase=2)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 959, in get_query_str
return self.run_query(client=client, phase=phase, **query_obj)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 1126, in run_query
client.timeseries(**qry)
File "/usr/lib/python2.7/site-packages/pydruid/client.py", line 167, in timeseries
return self._post(query)
File "/usr/lib/python2.7/site-packages/pydruid/client.py", line 484, in _post
res = urllib.request.urlopen(req)
File "/usr/lib64/python2.7/urllib2.py", line 154, in urlopen
return opener.open(url, data, timeout)
File "/usr/lib64/python2.7/urllib2.py", line 431, in open
response = self._open(req, data)
File "/usr/lib64/python2.7/urllib2.py", line 449, in _open
'_open', req)
File "/usr/lib64/python2.7/urllib2.py", line 409, in _call_chain
result = func(*args)
File "/usr/lib64/python2.7/urllib2.py", line 1244, in http_open
return self.do_open(httplib.HTTPConnection, req)
File "/usr/lib64/python2.7/urllib2.py", line 1214, in do_open
raise URLError(err)
URLError: <urlopen error [Errno -2] Name or service not known>
另见3.图
知道什么是问题吗?
我通过 NiFi 为 Kafka 提供数据,然后我将 kafka 源连接到 SAM 中的德鲁伊目标。
谢谢!
图片
图片
图片
超集中没有数据
似乎 Superset 在连接到您的 Broker-Node 时遇到了问题。检查集群运行状况。特别是 Broker- 和 Coordinator 节点日志。
问题已解决,
问题是代理主机未在超级集 UI 的集群配置中定义。
我将其设置为值:localhost
现在开始了 运行.
我正在使用 Kafka、Druid 和 SuperSet 测试数据流。
我在 Druid 中有一些数据(参见 1.picutre)。
之后我可以通过选项 "Refresh Druid metadata" 在 Superset 中生成 Druid 数据源(参见 2.pic) 问题是当我想查询数据时,我收到此错误消息:
URLError: <urlopen error [Errno -2] Name or service not known>
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/superset/viz.py", line 329, in get_df_payload
df = self.get_df(query_obj)
File "/usr/lib/python2.7/site-packages/superset/viz.py", line 142, in get_df
self.results = self.datasource.query(query_obj)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 1238, in query
client=client, query_obj=query_obj, phase=2)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 959, in get_query_str
return self.run_query(client=client, phase=phase, **query_obj)
File "/usr/lib/python2.7/site-packages/superset/connectors/druid/models.py", line 1126, in run_query
client.timeseries(**qry)
File "/usr/lib/python2.7/site-packages/pydruid/client.py", line 167, in timeseries
return self._post(query)
File "/usr/lib/python2.7/site-packages/pydruid/client.py", line 484, in _post
res = urllib.request.urlopen(req)
File "/usr/lib64/python2.7/urllib2.py", line 154, in urlopen
return opener.open(url, data, timeout)
File "/usr/lib64/python2.7/urllib2.py", line 431, in open
response = self._open(req, data)
File "/usr/lib64/python2.7/urllib2.py", line 449, in _open
'_open', req)
File "/usr/lib64/python2.7/urllib2.py", line 409, in _call_chain
result = func(*args)
File "/usr/lib64/python2.7/urllib2.py", line 1244, in http_open
return self.do_open(httplib.HTTPConnection, req)
File "/usr/lib64/python2.7/urllib2.py", line 1214, in do_open
raise URLError(err)
URLError: <urlopen error [Errno -2] Name or service not known>
另见3.图
知道什么是问题吗?
我通过 NiFi 为 Kafka 提供数据,然后我将 kafka 源连接到 SAM 中的德鲁伊目标。
谢谢!
图片
图片
图片
超集中没有数据
似乎 Superset 在连接到您的 Broker-Node 时遇到了问题。检查集群运行状况。特别是 Broker- 和 Coordinator 节点日志。
问题已解决, 问题是代理主机未在超级集 UI 的集群配置中定义。 我将其设置为值:localhost 现在开始了 运行.