Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
D
dataease
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
zhu
dataease
Commits
f3fad450
提交
f3fad450
authored
4月 21, 2021
作者:
taojinlong
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
feat: 测试性能
上级
6b285d73
全部展开
隐藏空白字符变更
内嵌
并排
正在显示
5 个修改的文件
包含
113 行增加
和
34 行删除
+113
-34
CommonConfig.java
backend/src/main/java/io/dataease/config/CommonConfig.java
+9
-9
JdbcProvider.java
...in/java/io/dataease/datasource/provider/JdbcProvider.java
+18
-4
AppStartReadHBaseListener.java
.../java/io/dataease/listener/AppStartReadHBaseListener.java
+15
-15
ExtractDataService.java
.../java/io/dataease/service/dataset/ExtractDataService.java
+0
-0
SparkCalc.java
...nd/src/main/java/io/dataease/service/spark/SparkCalc.java
+71
-6
没有找到文件。
backend/src/main/java/io/dataease/config/CommonConfig.java
浏览文件 @
f3fad450
...
...
@@ -22,15 +22,15 @@ public class CommonConfig {
private
Environment
env
;
// 保存了配置文件的信息
private
static
String
root_path
=
"/opt/dataease/data/kettle/"
;
@Bean
@ConditionalOnMissingBean
public
org
.
apache
.
hadoop
.
conf
.
Configuration
configuration
()
{
org
.
apache
.
hadoop
.
conf
.
Configuration
configuration
=
new
org
.
apache
.
hadoop
.
conf
.
Configuration
();
configuration
.
set
(
"hbase.zookeeper.quorum"
,
env
.
getProperty
(
"hbase.zookeeper.quorum"
));
configuration
.
set
(
"hbase.zookeeper.property.clientPort"
,
env
.
getProperty
(
"hbase.zookeeper.property.clientPort"
));
configuration
.
set
(
"hbase.client.retries.number"
,
env
.
getProperty
(
"hbase.client.retries.number"
,
"1"
));
return
configuration
;
}
//
@Bean
//
@ConditionalOnMissingBean
//
public org.apache.hadoop.conf.Configuration configuration() {
//
org.apache.hadoop.conf.Configuration configuration = new org.apache.hadoop.conf.Configuration();
//
configuration.set("hbase.zookeeper.quorum", env.getProperty("hbase.zookeeper.quorum"));
//
configuration.set("hbase.zookeeper.property.clientPort", env.getProperty("hbase.zookeeper.property.clientPort"));
//
configuration.set("hbase.client.retries.number", env.getProperty("hbase.client.retries.number", "1"));
//
return configuration;
//
}
@Bean
@ConditionalOnMissingBean
...
...
backend/src/main/java/io/dataease/datasource/provider/JdbcProvider.java
浏览文件 @
f3fad450
...
...
@@ -7,6 +7,7 @@ import io.dataease.datasource.dto.MysqlConfigrationDTO;
import
io.dataease.datasource.dto.SqlServerConfigration
;
import
io.dataease.datasource.dto.TableFiled
;
import
io.dataease.datasource.request.DatasourceRequest
;
import
org.apache.arrow.util.VisibleForTesting
;
import
org.apache.commons.lang3.StringUtils
;
import
org.springframework.stereotype.Service
;
...
...
@@ -39,6 +40,23 @@ public class JdbcProvider extends DatasourceProvider {
return
list
;
}
@VisibleForTesting
public
void
exec
(
DatasourceRequest
datasourceRequest
)
throws
Exception
{
Connection
connection
=
null
;
try
{
connection
=
getConnectionFromPool
(
datasourceRequest
);
Statement
stat
=
connection
.
createStatement
();
stat
.
execute
(
datasourceRequest
.
getQuery
());
}
catch
(
SQLException
e
)
{
throw
new
Exception
(
"ERROR:"
+
e
.
getMessage
(),
e
);
}
catch
(
Exception
e
)
{
throw
new
Exception
(
"ERROR:"
+
e
.
getMessage
(),
e
);
}
finally
{
returnSource
(
connection
,
datasourceRequest
.
getDatasource
().
getId
());
}
}
@Override
public
ResultSet
getDataResultSet
(
DatasourceRequest
datasourceRequest
)
throws
Exception
{
ResultSet
rs
;
...
...
@@ -47,7 +65,6 @@ public class JdbcProvider extends DatasourceProvider {
connection
=
getConnectionFromPool
(
datasourceRequest
);
Statement
stat
=
connection
.
createStatement
();
rs
=
stat
.
executeQuery
(
datasourceRequest
.
getQuery
());
returnSource
(
connection
,
datasourceRequest
.
getDatasource
().
getId
());
}
catch
(
SQLException
e
)
{
throw
new
Exception
(
"ERROR:"
+
e
.
getMessage
(),
e
);
}
catch
(
Exception
e
)
{
...
...
@@ -66,7 +83,6 @@ public class JdbcProvider extends DatasourceProvider {
connection
=
getConnectionFromPool
(
datasourceRequest
);
Statement
stat
=
connection
.
createStatement
();
ResultSet
rs
=
stat
.
executeQuery
(
datasourceRequest
.
getQuery
()
+
MessageFormat
.
format
(
" LIMIT {0}, {1}"
,
(
datasourceRequest
.
getStartPage
()
-
1
)
*
datasourceRequest
.
getPageSize
(),
datasourceRequest
.
getPageSize
()));
returnSource
(
connection
,
datasourceRequest
.
getDatasource
().
getId
());
list
=
fetchResult
(
rs
);
}
catch
(
SQLException
e
)
{
throw
new
Exception
(
"ERROR:"
+
e
.
getMessage
(),
e
);
...
...
@@ -174,8 +190,6 @@ public class JdbcProvider extends DatasourceProvider {
return
list
;
}
;
@Override
public
void
test
(
DatasourceRequest
datasourceRequest
)
throws
Exception
{
String
queryStr
=
getTablesSql
(
datasourceRequest
);
...
...
backend/src/main/java/io/dataease/listener/AppStartReadHBaseListener.java
浏览文件 @
f3fad450
...
...
@@ -33,20 +33,20 @@ public class AppStartReadHBaseListener implements ApplicationListener<Applicatio
@Override
public
void
onApplicationEvent
(
ApplicationReadyEvent
applicationReadyEvent
)
{
System
.
out
.
println
(
"================= Read HBase start ================="
);
// 项目启动,从数据集中找到定时抽取的表,从HBase中读取放入缓存
DatasetTableExample
datasetTableExample
=
new
DatasetTableExample
();
datasetTableExample
.
createCriteria
().
andModeEqualTo
(
1
);
List
<
DatasetTable
>
datasetTables
=
datasetTableMapper
.
selectByExampleWithBLOBs
(
datasetTableExample
);
for
(
DatasetTable
table
:
datasetTables
)
{
// commonThreadPool.addTask(() -> {
try
{
List
<
DatasetTableField
>
fields
=
dataSetTableFieldsService
.
getFieldsByTableId
(
table
.
getId
());
sparkCalc
.
getHBaseDataAndCache
(
table
.
getId
(),
fields
);
}
catch
(
Exception
e
)
{
e
.
printStackTrace
();
}
// });
}
//
System.out.println("================= Read HBase start =================");
//
// 项目启动,从数据集中找到定时抽取的表,从HBase中读取放入缓存
//
DatasetTableExample datasetTableExample = new DatasetTableExample();
//
datasetTableExample.createCriteria().andModeEqualTo(1);
//
List<DatasetTable> datasetTables = datasetTableMapper.selectByExampleWithBLOBs(datasetTableExample);
//
for (DatasetTable table : datasetTables) {
//
//
commonThreadPool.addTask(() -> {
//
try {
//
List<DatasetTableField> fields = dataSetTableFieldsService.getFieldsByTableId(table.getId());
//
sparkCalc.getHBaseDataAndCache(table.getId(), fields);
//
} catch (Exception e) {
//
e.printStackTrace();
//
}
//
//
});
//
}
}
}
backend/src/main/java/io/dataease/service/dataset/ExtractDataService.java
浏览文件 @
f3fad450
差异被折叠。
点击展开。
backend/src/main/java/io/dataease/service/spark/SparkCalc.java
浏览文件 @
f3fad450
...
...
@@ -41,6 +41,7 @@ import java.util.List;
@Service
public
class
SparkCalc
{
private
static
String
column_family
=
"dataease"
;
private
static
String
data_path
=
"/opt/dataease/data/db/"
;
@Resource
private
Environment
env
;
// 保存了配置文件的信息
...
...
@@ -54,12 +55,13 @@ public class SparkCalc {
sqlContext
.
setConf
(
"spark.sql.shuffle.partitions"
,
env
.
getProperty
(
"spark.sql.shuffle.partitions"
,
"1"
));
sqlContext
.
setConf
(
"spark.default.parallelism"
,
env
.
getProperty
(
"spark.default.parallelism"
,
"1"
));
Dataset
<
Row
>
dataFrame
=
CacheUtil
.
getInstance
().
getCacheData
(
hTable
);
if
(
ObjectUtils
.
isEmpty
(
dataFrame
))
{
dataFrame
=
getHBaseDataAndCache
(
sparkContext
,
sqlContext
,
hTable
,
fields
);
}
Dataset
<
Row
>
dataFrame
=
getData
(
sparkContext
,
sqlContext
,
hTable
,
fields
);
// Dataset<Row> dataFrame = CacheUtil.getInstance().getCacheData(hTable);
// if (ObjectUtils.isEmpty(dataFrame)) {
// dataFrame = getData(sparkContext, sqlContext, hTable, fields);
// }
dataFrame
.
createOrReplaceTempView
(
tmpTable
);
dataFrame
.
createOrReplaceTempView
(
tmpTable
);
Dataset
<
Row
>
sql
=
sqlContext
.
sql
(
getSQL
(
xAxis
,
yAxis
,
tmpTable
,
requestList
));
// transform
List
<
String
[]>
data
=
new
ArrayList
<>();
...
...
@@ -86,6 +88,69 @@ public class SparkCalc {
return
getHBaseDataAndCache
(
sparkContext
,
sqlContext
,
hTable
,
fields
);
}
public
Dataset
<
Row
>
getData
(
JavaSparkContext
sparkContext
,
SQLContext
sqlContext
,
String
tableId
,
List
<
DatasetTableField
>
fields
)
throws
Exception
{
fields
.
sort
((
o1
,
o2
)
->
{
if
(
o1
.
getOriginName
()
==
null
)
{
return
-
1
;
}
if
(
o2
.
getOriginName
()
==
null
)
{
return
1
;
}
return
o1
.
getOriginName
().
compareTo
(
o2
.
getOriginName
());
});
JavaRDD
<
String
>
pairRDD
=
sparkContext
.
textFile
(
data_path
+
tableId
+
".txt"
);
JavaRDD
<
Row
>
rdd
=
pairRDD
.
mapPartitions
(
(
FlatMapFunction
<
java
.
util
.
Iterator
<
String
>,
Row
>)
tuple2Iterator
->
{
List
<
Row
>
iterator
=
new
ArrayList
<>();
while
(
tuple2Iterator
.
hasNext
())
{
String
[]
items
=
tuple2Iterator
.
next
().
split
(
";"
);
List
<
Object
>
list
=
new
ArrayList
<>();
for
(
int
i
=
0
;
i
<
items
.
length
;
i
++){
String
l
=
items
[
i
];
DatasetTableField
x
=
fields
.
get
(
i
);
if
(
x
.
getDeType
()
==
0
||
x
.
getDeType
()
==
1
)
{
list
.
add
(
l
);
}
else
if
(
x
.
getDeType
()
==
2
)
{
if
(
StringUtils
.
isEmpty
(
l
))
{
l
=
"0"
;
}
if
(
StringUtils
.
equalsIgnoreCase
(
l
,
"Y"
))
{
l
=
"1"
;
}
if
(
StringUtils
.
equalsIgnoreCase
(
l
,
"N"
))
{
l
=
"0"
;
}
list
.
add
(
Long
.
valueOf
(
l
));
}
else
if
(
x
.
getDeType
()
==
3
)
{
if
(
StringUtils
.
isEmpty
(
l
))
{
l
=
"0.0"
;
}
list
.
add
(
Double
.
valueOf
(
l
));
}
}
iterator
.
add
(
RowFactory
.
create
(
list
.
toArray
()));
}
return
iterator
.
iterator
();
});
List
<
StructField
>
structFields
=
new
ArrayList
<>();
// struct顺序要与rdd顺序一致
fields
.
forEach
(
x
->
{
if
(
x
.
getDeType
()
==
0
||
x
.
getDeType
()
==
1
)
{
structFields
.
add
(
DataTypes
.
createStructField
(
x
.
getOriginName
(),
DataTypes
.
StringType
,
true
));
}
else
if
(
x
.
getDeType
()
==
2
)
{
structFields
.
add
(
DataTypes
.
createStructField
(
x
.
getOriginName
(),
DataTypes
.
LongType
,
true
));
}
else
if
(
x
.
getDeType
()
==
3
)
{
structFields
.
add
(
DataTypes
.
createStructField
(
x
.
getOriginName
(),
DataTypes
.
DoubleType
,
true
));
}
});
StructType
structType
=
DataTypes
.
createStructType
(
structFields
);
Dataset
<
Row
>
dataFrame
=
sqlContext
.
createDataFrame
(
rdd
,
structType
);
return
dataFrame
;
}
public
Dataset
<
Row
>
getHBaseDataAndCache
(
JavaSparkContext
sparkContext
,
SQLContext
sqlContext
,
String
hTable
,
List
<
DatasetTableField
>
fields
)
throws
Exception
{
Scan
scan
=
new
Scan
();
scan
.
addFamily
(
Bytes
.
toBytes
(
column_family
));
...
...
@@ -145,7 +210,7 @@ public class SparkCalc {
StructType
structType
=
DataTypes
.
createStructType
(
structFields
);
Dataset
<
Row
>
dataFrame
=
sqlContext
.
createDataFrame
(
rdd
,
structType
).
persist
(
StorageLevel
.
MEMORY_AND_DISK_SER
());
CacheUtil
.
getInstance
().
addCacheData
(
hTable
,
dataFrame
);
//
CacheUtil.getInstance().addCacheData(hTable, dataFrame);
dataFrame
.
count
();
return
dataFrame
;
}
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论