
A. Liang et al.
Fig. 1 Four dissimilar trajectories with a similar portion (the part in the
black box)
Existing solutions to the sub-trajectory clustering problem
follow the first-segment-then-cluster paradigm. The effec-
tiveness of the clustering result heavily depends on the
segmentation step, which is a crucial diff erence from the
trajectory clustering problem. Existing solutions can be clas-
sified into two categories. The first one regards segmentation
and clustering as two independent procedures. It segments
the t rajectories into sub-trajectories according to spatial fea-
tures and then clusters these sub-trajectories. For instance,
Lee et al. [18] proposed TRACLUS, which segments trajec-
tories into line segments based on the minimum description
length principle and then clusters line segments by applying
DBSCAN. The second one segments trajectories by taking
the clustering quality into account. For example, Pelekis et
al. [25] segment trajectories based on the local density cri-
terion and then conduct clustering according to the density.
These methods usually rely on complicated hand-crafted seg-
mentation rules tailored to specific applications, making the
clustering result sensitive to these rules and limiting t heir
generality.
To address these issues, we consider utilizing the clus-
tering quality to guide the segmentation in a data-driven
manner instead of using hand-crafted segmentation rules.
A straightforward approach is enumerating all segmentation
schemes and choosing the one with the best clustering qual-
ity. However, this method is computationally expensive and
is thus impractical for real-world applications. Therefore,
we propose an efficient and effective reinforcement learn-
ing (RL)-based framework that leverages clustering quality
to guide segmentation. It allows segmentation and clustering
components to cooperate closely, mutually enhancing each
other and achieving superior clustering results.
Specifically, we treat the segmentation procedure as a
sequential decision process, i.e., it sequentially scans the
trajectory points and decides whether to segment at that
point, which can be modeled as a Markov decision process
(MDP) [26]. The crucial challenge of the MDP formulation
is to define the state and reward, which significantly affect
the performance of the learned model, i.e., the RL agent.
Regarding the design of the state, the intuitive idea is to
include more features to reflect the environment. However,
experimental results indicate that such an approach may lead
to much longer training time without noticeable improve-
ments i n the model’s performance. To solve this problem, we
propose a representative set of features that can effectively
summarize the sub-trajectory clustering characteristics while
maintaining acceptable training time. Designing an appropri-
ate reward is vital to the training process as it encourages the
RL agent to choose good actions to boost performance. We
define the reward based on the evaluation metrics of clus-
tering quality, which allows us to leverage the clustering
quality to guide the trajectory segmentation, thus achiev-
ing the data-driven trajectory segmentation. We employ the
Deep-Q-Network (DQN) method [24] to learn the policy for
the MDP in our trajectory segmentation problem. Based on
the learned policy, we propose a RL-based Sub-Trajectory
Clustering algorithm called RLSTC. Our algorithm is data-
driven and can adapt to different dynamics of the underlying
trajectories. Moreover, it is a general solution that can be
applied to various trajectory similarity measurements and
application scenarios.
To further improve efficiency, we propose several opti-
mizations to decrease the cost of trajectory distance computa-
tion. First, we implement a preprocessing step that simplifies
trajectories by reserving only the critical points. Second, we
apply an approximate method to compute trajectory similar-
ity in linear time. Finally, we utilize an incremental similarity
computation strategy to avoid recomputing the distance from
scratch.
In summary, our main contributions are as follows:
• We propose a novel RL-based framework for the sub-
trajectory clustering problem, which is the first RL-based
solution to the best of our knowledge.
• We creatively model the procedure of trajectory seg-
mentation as an MDP, especially defining the state and
reward of MDP by considering the distinct features
of the sub-trajectory clustering. This method achieves
data-driven trajectory segmentation without relying on
complex hand-crafted rules.
• We have developed several optimizations to enhance the
efficiency of our solution.
• We conduct extensive experiments on various real-world
trajectory datasets. The experimental results demonstrate
that our RL-based algorithm outperforms state-of-the-art
methods in terms of both effectiveness and efficiency.
The remainder of the paper is as follows: In Sect. 2,we
present preliminaries and our problem statement. Section 3
discusses the preprocessing algorithm. In Sect. 4, we detail
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
文档被以下合辑收录
评论