Java 类名:com.alibaba.alink.operator.batch.clustering.GroupDbscanModelBatchOp
Python 类名:GroupDbscanModelBatchOp
DBSCAN,Density-Based Spatial Clustering of Applications with Noise,是一个比较有代表性的基于密度的聚类算法。与划分和层次聚类方法不同,它将簇定义为密度相连的点的最大集合,能够把具有足够高密度的区域划分为簇,并可在噪声的空间数据库中发现任意形状的聚类。

| 参数名称 | 参数描述 | 说明 |
|---|---|---|
| EUCLIDEAN | |
欧式距离 |
| COSINE | |
夹角余弦距离 |
| CITYBLOCK | |
城市街区距离,也称曼哈顿距离 |
| 名称 | 中文名称 | 描述 | 类型 | 是否必须? | 取值范围 | 默认值 |
|---|---|---|---|---|---|---|
| epsilon | 邻域距离阈值 | 邻域距离阈值 | Double | ✓ | ||
| featureCols | 特征列名 | 特征列名,必选 | String[] | ✓ | 所选列类型为 [BIGDECIMAL, BIGINTEGER, BYTE, DOUBLE, FLOAT, INTEGER, LONG, SHORT] | |
| groupCols | 分组列名,多列 | 分组列名,多列,必选 | String[] | ✓ | ||
| minPoints | 邻域中样本个数的阈值 | 邻域中样本个数的阈值 | Integer | ✓ | ||
| predictionCol | 预测结果列名 | 预测结果列名 | String | ✓ | ||
| distanceType | 距离度量方式 | 聚类使用的距离类型 | String | “EUCLIDEAN”, “COSINE”, “CITYBLOCK”, “HAVERSINE”, “JACCARD” | “EUCLIDEAN” | |
| groupMaxSamples | 每个分组的最大样本数 | 每个分组的最大样本数 | Integer | 2147483647 | ||
| skip | 每个分组超过最大样本数时,是否跳过 | 每个分组超过最大样本数时,是否跳过 | Boolean | false |
from pyalink.alink import *
import pandas as pd
useLocalEnv(1)
df = pd.DataFrame([
[0, "id_1", 2.0, 3.0],
[0, "id_2", 2.1, 3.1],
[0, "id_18", 2.4, 3.2],
[0, "id_15", 2.8, 3.2],
[0, "id_12", 2.1, 3.1],
[0, "id_3", 200.1, 300.1],
[0, "id_4", 200.2, 300.2],
[0, "id_8", 200.6, 300.6],
[1, "id_5", 200.3, 300.3],
[1, "id_6", 200.4, 300.4],
[1, "id_7", 200.5, 300.5],
[1, "id_16", 300., 300.2],
[1, "id_9", 2.1, 3.1],
[1, "id_10", 2.2, 3.2],
[1, "id_11", 2.3, 3.3],
[1, "id_13", 2.4, 3.4],
[1, "id_14", 2.5, 3.5],
[1, "id_17", 2.6, 3.6],
[1, "id_19", 2.7, 3.7],
[1, "id_20", 2.8, 3.8],
[1, "id_21", 2.9, 3.9],
[2, "id_20", 2.8, 3.8]])
source = BatchOperator.fromDataframe(df, schemaStr='group string, id string, c1 double, c2 double')
groupDbscan = GroupDbscanModelBatchOp()\
.setGroupCols(["group"])\
.setFeatureCols(["c1", "c2"])\
.setMinPoints(4)\
.setEpsilon(0.6)\
.linkFrom(source)
groupDbscan.print()
import org.apache.flink.types.Row;
import com.alibaba.alink.operator.batch.BatchOperator;
import com.alibaba.alink.operator.batch.source.MemSourceBatchOp;
import com.alibaba.alink.operator.batch.clustering.GroupDbscanModelBatchOp;
import org.junit.Test;
import java.util.Arrays;
import java.util.List;
public class GroupDbscanModelBatchOpTest {
@Test
public void testGroupDbscanModelBatchOp() throws Exception {
List<Row> trainData = Arrays.asList(
Row.of(0, "id_1", 2.0, 3.0),
Row.of(0, "id_2", 2.1, 3.1),
Row.of(0, "id_18", 2.4, 3.2),
Row.of(0, "id_15", 2.8, 3.2),
Row.of(0, "id_12", 2.1, 3.1),
Row.of(0, "id_3", 200.1, 300.1),
Row.of(0, "id_4", 200.2, 300.2),
Row.of(0, "id_8", 200.6, 300.6),
Row.of(1, "id_5", 200.3, 300.3),
Row.of(1, "id_6", 200.4, 300.4),
Row.of(1, "id_7", 200.5, 300.5),
Row.of(1, "id_16", 300., 300.2),
Row.of(1, "id_9", 2.1, 3.1),
Row.of(1, "id_10", 2.2, 3.2),
Row.of(1, "id_11", 2.3, 3.3),
Row.of(1, "id_13", 2.4, 3.4),
Row.of(1, "id_14", 2.5, 3.5),
Row.of(1, "id_17", 2.6, 3.6),
Row.of(1, "id_19", 2.7, 3.7),
Row.of(1, "id_20", 2.8, 3.8),
Row.of(1, "id_21", 2.9, 3.9),
Row.of(2, "id_20", 2.8, 3.8)
);
MemSourceBatchOp inputOp = new MemSourceBatchOp(trainData,
new String[] {"group", "id", "c1", "c2"});
GroupDbscanModelBatchOp op = new GroupDbscanModelBatchOp()
.setGroupCols("group")
.setFeatureCols("c1", "c2")
.setMinPoints(4)
.setEpsilon(0.6)
.linkFrom(inputOp);
op.print();
}
}
| group | cluster_id | count | c1 | c2 |
|---|---|---|---|---|
| 1 | 0 | 9 | 2.5000 | 3.5000 |
| 0 | 0 | 5 | 2.2800 | 3.1200 |