<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">An Efficient Framework for the Clustering of Human Activity Data using Kernelized Robust Covariance Descriptors</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Guntru</forename><forename type="middle">Prasanth</forename><surname>Kumar</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">National Institute of Technology Calicut</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><roleName>M. S. Subodh Raj</roleName><forename type="first">Sudhish</forename><forename type="middle">N</forename><surname>George</surname></persName>
							<email>sudhish@nitc.ac.in</email>
							<affiliation key="aff0">
								<orgName type="institution">National Institute of Technology Calicut</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">An Efficient Framework for the Clustering of Human Activity Data using Kernelized Robust Covariance Descriptors</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">62D02F59B6E761F954BE05643235759A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T19:50+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Human activity data</term>
					<term>Kernelized covariance descriptors</term>
					<term>Temporal Laplacian regularization</term>
					<term>Temporal subspace clustering</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, a new method for the efficient clustering of human activity data is proposed. Unlike the traditional human activity clustering approaches, our method relies on the skeletal data recorded with the help of motion capture (mocap) systems to achieve the goal. The proposed method is structured around the kernel-based robust covariance descriptor. By introducing a data re-framing technique that efficiently utilizes the temporal properties of the human activity data, we have alleviated the data redundancy and insufficiency issues associated with action sequences. The optimization model developed encompasses the combined benefits of low-rank representation and least square regression. The formulation is strengthened by incorporating the temporal dependency of the human activity sequences with the help of a temporal Laplacian regularizer. With the proposed algorithm, a representation matrix is learned from the raw data, which is then used to perform subspace clustering. Experiments conducted on multiple human activity datasets reveal the ability of the proposed method to achieve better clustering results compared to state-of-the-art counterparts.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Human activity recognition (HAR) from action sequences remains a challenging research topic in computer vision due to its multifaceted applications <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. HAR finds application in visual surveillance, healthcare, human-machine interface, video retrieval, and entertainment industry, to name a few <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. Traditional approaches to HAR use RGB video sequences as the input. Handcrafted features are later extracted from the video sequences for the purpose of activity recognition <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. Because of the high-dimensional nature of the video sequences, high computational complexity is often associated with such HAR approaches <ref type="bibr" target="#b0">[1]</ref>. Later, sensor-based HAR gained popularity. The focus of such methods were on the data obtained from sensors such as accelerometer and gyroscopes. In such cases, the subject itself needs to be in possession of the sensor so that the movements of the human body can be recorded. This along with the influence of noise acts as a limiting factor in sensor based HAR approaches <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b0">1,</ref><ref type="bibr" target="#b5">6]</ref>. With the evolution of mocap systems, new modalities were introduced to represent the human activity information. Such modalities include the motion depth maps and the skeletal representations <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref>. With skeletal representations, time series of 3D joint positions of the human body are recorded by the mocap systems <ref type="bibr" target="#b8">[9]</ref>. The spatio-temporal quality of such recordings are superior that they find application in multiple domains including gait analysis, medical rehabilitation, and computer animations <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b2">3]</ref>.</p><p>The common approaches to HAR with mocap data involve supervised learning methods. Though they guarantee good results, such methods are often susceptible to missing sample issues. Further, the requirement of a huge clean dataset for initial learning of the system poses a serious bottleneck to supervised HAR approaches <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b9">10]</ref>. This paves a foundation for the need of having unsupervised HAR strategies. Though in unsupervised methods the aforementioned challenges faced by supervised methods are alleviated, they encounter other limitations. The main challenge in unsupervised HAR is posed by the fact that the task needs to be performed in a robust and accurate manner without any prior knowledge about the data samples <ref type="bibr" target="#b10">[11]</ref>.</p><p>Recent studies showcase the ability of subspace clustering algorithms in dealing with highdimensional data clustering problems <ref type="bibr" target="#b11">[12]</ref>. The key idea in subspace clustering approach is to identify multiple low-dimensional subspaces from which the data originates <ref type="bibr" target="#b12">[13]</ref>. The subspaces so identified house a cluster of data. This approach is commonly termed as the Union of Subspaces (UoS) model <ref type="bibr" target="#b13">[14]</ref>. The popularity of subspace clustering approach has increased with the introduction of sparse subspace clustering approach in which a sparsity constraint is imposed on the coefficients in order to learn a sparse representation of the raw data <ref type="bibr" target="#b13">[14]</ref>. Low-rank representation learning (LRR) <ref type="bibr" target="#b14">[15]</ref> is another technique used with subspace clustering wherein the global structure of the data is considered while learning the coefficients. In LRR based approaches, a given dictionary is utilized to learn a low-rank representation of the data samples. The clustering results obtained with such low-rank representations are usually better <ref type="bibr" target="#b15">[16]</ref>. Least square regression (LSR) <ref type="bibr" target="#b16">[17]</ref> based subspace clustering in which the grouping of data samples are performed with the help of Frobenius norm operator is another promising approach in subspace clustering. The aforementioned approaches when used independently will not be suitable for HAR as they do not consider the time series information associated with the skeletal data. As a solution to this problem we have developed an approach called the time series activity clustering (TSAC) which utilizes the combined advantages of LRR and LSR. We have incorporated a kernel-based robust covariance descriptor to extract features out of the raw input data with the aim of exploring the non-linearity present in the data. Further, the temporal Laplacian regularizer is employed to capture the temporal dependency among the data samples.</p><p>The following are the main contributions of this work:</p><p>1. An unsupervised optimization model with improved performance is formulated for the clustering of human activity data. To effectively utilize the temporal dependencies of human activity sequences, a temporal Laplacian regularizer is introduced in the proposed model. 2. We have blended the LRR and LSR based subspace clustering approaches to achieve better clustering results. A clean dictionary is learned along with the representation matrix in The rest of the work is outlined as follows. The proposed method, the problem formulation, and the solutions obtained are presented in Section 2. Section 3 explains the experimental validation done using the proposed method. Finally, conclusions are drawn in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Proposed Method</head><p>In this section, we give a detailed outline of the proposed TSAC approach. The workflow of the proposed approach is shown in Fig. <ref type="figure" target="#fig_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Data Re-framing</head><p>A sequence of human action is a collection of action timestamps evolved over a period of time. Two important observations can be made with reference to a given collection of action sequences as mentioned below:</p><p>1. The number of timestamps corresponding to each of the action sequences may not necessarily be the same. This disparity often appears as a bottleneck in generalizing any algorithm dealing with human activity recognition. 2. As the number of timestamps in an action sequence increases, it introduces additional computation overhead. That will lead to a rise in demand for resource utilization.</p><p>The aforementioned challenges can be addressed by standardizing the number of timestamps for all the action sequences under consideration. If the number of timestamps is standardized to be 'N ', then data pruning methods need to be employed on action sequences having more number of timestamps and data augmentation needs to be performed on action sequences having lesser number of timestamps. The temporal smoothness property of human action sequences can be conveniently utilized to achieve this goal Human action sequences are temporally highly correlated. As far as activity recognition is concerned, this correlation leads to redundant information. Often, the complete set of frames of a recorded action sequence is not essential to perform activity recognition. Thus, we introduce a pruning technique, termed succession pruning, in which the alternative frames of the action sequence are removed to eliminate redundancy while maintaining the temporal properties of the action sequence. That will drastically reduce the amount of data to be processed and will also result in reduced computation overhead and resource utilization. Whereas in action sequences experiencing insufficiency of timestamps, we perform timestamp augmentation. In this process, the trailing end of the action sequence gets augmented with the terminal timestamps of the same action sequence. That is in line with the temporal smoothness property of the human action data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Feature Matrix Generation using Kernel-based Robust Covariance Descriptor</head><p>Human actions are represented in the form of skeletal structures with the help of modern motion capture systems. Each timestamp of an action sequence can be represented as a collection of 'n' joints. Thus, the time stamp 't' of an action sequence can be represented as e(𝑡) ∈ R 3×𝑛 of 3D positions {e 1 (𝑡), . . . , e 𝑛 (𝑡)}. The 3D coordinates of the i 𝑡ℎ joint of the skeletal structure corresponding to the t 𝑡ℎ timestamp can be denoted as, e 𝑖 (𝑡) ∈ [𝑥 𝑖 (𝑡), 𝑦 𝑖 (𝑡), 𝑧 𝑖 (𝑡)] ⊤ .</p><p>Once the data re-framing process is completed, the raw action sequences with fixed number of timestamps are represented in the form of a feature matrix. It involves a two step process. In the first step, we use the concept of covariance to obtain the covariance descriptor of each action sequence. The use of covariance will help us to capture the changes pertaining to each joint of the skeletal structure <ref type="bibr" target="#b17">[18]</ref>. If 𝜇 represents the temporal average of the timestamps of an action sequence, then the corresponding action sequence can be represented in the form of a covariance matrix as shown below.</p><formula xml:id="formula_0">Ψ = 1 𝑁 − 1 𝑁 ∑︁ 𝑡=1 [e(𝑡) − 𝜇][e(𝑡) − 𝜇] ⊤<label>(1)</label></formula><p>This process is repeated for each action sequence, resulting in a unique covariance descriptor for each input sequence.</p><p>Although covariance descriptors finds application in multiple domains, they cannot capture non-linearity present in the data. For making the feature matrix robust, different approaches are adopted to incorporate additional statistical information along with the covariance descriptor. This includes the entropy based approaches, the mutual information based approaches, and the kernel-based approaches. Among others, the kernel-based approaches have been used to simulate more complex models. The use of kernels improves the descriptive power of the covariance matrices. The work proposed by Cavazza et al. <ref type="bibr" target="#b17">[18]</ref> showcases the benefits of using kernels in works related to human activity recognition. Motivated by this observation, we modify the expression given in Eq. ( <ref type="formula" target="#formula_0">1</ref>) to incorporate the kernel function and obtain the following robust covariance descriptor.</p><formula xml:id="formula_1">Ψ = 1 𝑁 − 1 𝑁 ∑︁ 𝑡=1 [︀ 𝒦(e(𝑡)) − 𝜇 𝜅 ]︀[︀ 𝒦(e(𝑡)) − 𝜇 𝜅 ]︀ ⊤<label>(2)</label></formula><p>Here, 𝒦(.) represents the kernel function and 𝜇 𝜅 is the temporal average of the kernel entries. The choice of kernel function is application specific. We have used two kernel functions namely the polynomial kernel and the exponential kernel, out of which the later one have produced promising results. The exponential kernel is defined as:</p><p>𝒦(e(𝑡)) = exp {︁ e(𝑡)</p><formula xml:id="formula_2">(𝜎 + 𝑏) 2 }︁ (<label>3</label></formula><formula xml:id="formula_3">)</formula><p>where 𝑏 &gt; 0 and 𝜎 is the kernel bandwidth.</p><p>After obtaining the robust covariance descriptors for each action sequence, in the second step we generate the feature matrix. The covariance descriptor Ψ contains redundant information as it is symmetric along the main diagonal. In order to reduce the amount of data to be processed, we vectorize each covariance descriptor by retaining the upper triangular values alone. Later, we stack (as columns) each of the vectors so obtained to form the feature matrix X ∈ R 𝑝×𝑘 . Here, '𝑝' is the length of the individual vectors and 'k' is the total number of action sequences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Temporal Subspace Clustering (TSC) with Laplacian Regularization</head><p>Subspace clustering is performed by generating a representation matrix out of the feature matrix in state-of-the-art approaches. This is accomplished by using the self representation property of the feature matrix, resulting in the computation of a set of unique coefficients Y ∈ R 𝑘×𝑘 for the feature matrix X ∈ R 𝑝×𝑘 . This can be mathematically expressed as X = XY. The limitation of such an approach is that they tend to produce sub-optimal results if the sampling done is not sufficient. Instead of using the self representation property, if we learn an efficient dictionary W ∈ R 𝑝×𝑘 , we can overcome the aforementioned problem. Given a dictionary W, the set of data samples X can be expressed as X ≈ WY. The set of coefficients Y can be efficiently obtained in an iterative manner by utilizing the underlying low-rank nature of Y. The low-rank property of Y can be capture with the help of LRR <ref type="bibr" target="#b14">[15]</ref>. Since the rank minimization problem is inherently NP-hard, nuclear norm can be used as a substitute for rank minimization <ref type="bibr" target="#b16">[17]</ref>. It is also important to obtain the intra-cluster correlation among the data samples to obtain better clustering results. To this end, we can use the principle of LSR <ref type="bibr" target="#b16">[17]</ref>. By yielding the concepts of LRR and LSR we formulate an optimization problem as follows.</p><formula xml:id="formula_4">min Y,W 1 2 ‖X − WY‖ 2 F + 𝜆 1 2 ‖Y‖ 2 F + 𝜆 2 ‖Y‖ * s.t. Y ≥ 0, W ≥ 0 (4)</formula><p>where, the term Manifold regularization methods are efficient in incorporating the temporal dependency of data in the problem formulation <ref type="bibr" target="#b18">[19]</ref>. Thus, by modifying the general Laplacian regularizer, which captures the spatial dependency of data, we have developed a temporal Laplacian regularizer ℒ(•) as our interest is in the temporal information of the action sequences.</p><p>For a representation matrix Y, the temporal Laplacian regularization function can be defined as <ref type="bibr" target="#b19">[20]</ref>:</p><formula xml:id="formula_5">ℒ(Y) = 1 2 ∑︁ 𝑖 ∑︁ 𝑗 𝑧 𝑖𝑗 ‖𝑦 𝑖 − 𝑦 𝑗 ‖ 2 2 = tr (YL T Y ⊤ ),<label>(5)</label></formula><p>where</p><formula xml:id="formula_6">L T = ̃︁ W − Z is a temporal Laplacian matrix, ̃︁ W ii = ∑︀ 𝑚 𝑗=1 𝑧 𝑖𝑗 ,</formula><p>and Z is a weight matrix that finds the successive similarities in X.</p><p>Each element of Z is found as <ref type="bibr" target="#b20">[21]</ref>,</p><formula xml:id="formula_7">𝑧 𝑖𝑗 = {︃ 1 for |𝑖 − 𝑗| ≤ 𝛾 2 0 otherwise (<label>6</label></formula><formula xml:id="formula_8">)</formula><p>where 𝛾 denotes empirically defined threshold value.</p><p>With the introduction of ℒ(Y), Eq. ( <ref type="formula">4</ref>) is modified as,</p><formula xml:id="formula_9">min Y,W 1 2 ‖X − WY‖ 2 F + 𝜆 1 2 ‖Y‖ 2 F + 𝜆 2 ‖Y‖ * + 𝜆 3 ℒ(Y) s.t. Y ≥ 0, W ≥ 0 (7)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Solution</head><p>The optimization problem given in Eq. ( <ref type="formula">7</ref>) can be solved using the ADMM approach under the ALM framework. ADMM finds solution for the unconstrained optimization problem by splitting the problem into multiple sub-problems. As a first step, we will introduce three auxiliary variables E, F, and G to decouple the terms present in the formulation. This results in a formulation as shown below.</p><p>min</p><formula xml:id="formula_10">Y,W,E,F,G 1 2 ‖X − WY‖ 2 F + 𝜆 1 2 ‖E‖ 2 F + 𝜆 2 ‖F‖ * + 𝜆 3 ℒ(G) s.t. Y = E, Y = F, Y = G, Y ≥ 0, W ≥ 0<label>(8)</label></formula><p>The Augmented Lagrangian corresponding to Eq. ( <ref type="formula" target="#formula_10">8</ref>) is given as:</p><formula xml:id="formula_11">L(E, F, G, Y, W) = 1 2 ‖X − WY‖ 2 F + 𝜆 1 2 ‖E‖ 2 F + 𝜆 2 ‖F‖ * + 𝜆 3 tr(GL T G ⊤ ) + ⟨Φ 1 , Y − E⟩ + ⟨Φ 2 , Y − F⟩ + ⟨Φ 3 , Y − G⟩ + 𝛽 2 (‖Y − E‖ 2 F + ‖Y − F‖ 2 F + ‖Y − G‖ 2 F )<label>(9)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.1.">Updating E:</head><p>The update expression for E is obtained by solving the following sub-problem.</p><formula xml:id="formula_12">E [𝑙+1] = argmin E 𝜆 1 ‖E‖ 2 F + ⟨Φ 1 , Y − E⟩ + 𝛽 2 ‖Y − E‖ 2 F (10)</formula><p>By differentiating Eq. ( <ref type="formula">10</ref>) with respect to E and equating it to zero, the E update is given as,</p><formula xml:id="formula_13">E [𝑙+1] = 1 𝜆 1 + 𝛽 (︀ Φ [𝑙] 1 + 𝛽Y [𝑙] )︀<label>(11)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.2.">Updating F:</head><p>The F sub-problem is given as,</p><formula xml:id="formula_14">F [𝑙+1] = argmin F 𝜆 2 ‖F‖ * + ⟨𝜑 2 , Y − F⟩ + 𝛽 2 ‖Y − F‖ 2 F (<label>12</label></formula><formula xml:id="formula_15">)</formula><p>The update expression for F is found using the singular value thresholding (SVT) operator as follows <ref type="bibr" target="#b15">[16]</ref>,</p><formula xml:id="formula_16">F [𝑙+1] = SVT 𝜆 2 𝛽 [︃ Y [𝑙] + Φ [𝑙] 2 𝛽 ]︃<label>(13)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.3.">Updating G:</head><p>The update expression for G is obtained by solving the following sub-problem.</p><formula xml:id="formula_17">G [𝑙+1] = argmin G 𝜆 3 tr(GL T G ⊤ ) + ⟨Φ 3 , Y − G⟩ + 𝛽 2 ‖Y − G‖ 2 F (<label>14</label></formula><formula xml:id="formula_18">)</formula><p>By differentiating Eq. ( <ref type="formula" target="#formula_17">14</ref>) with respect to G and equating it to zero, the G update is given as,</p><formula xml:id="formula_19">G [𝑙+1] = (︀ Φ [𝑙] 3 + 𝛽Y [𝑙] )︀(︀ 𝜆 3 (L T + L T ⊤ ) + 𝛽I )︀ −1<label>(15)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.4.">Updating Y:</head><p>By solving the following sub-problem, the update expression for Y can be obtained.</p><formula xml:id="formula_20">Y [𝑙+1] = argmin Y 1 2 ‖X − WY‖ 2 F + ⟨Φ 1 , Y − E⟩ + ⟨Φ 2 , Y − F⟩ + ⟨Φ 3 , Y − G⟩ + 𝛽 2 (︀ ‖Y − E‖ 2 F + ‖Y − F‖ 2 F + ‖Y − G‖ 2 F )︀<label>(16)</label></formula><p>By equating the gradient of Eq. ( <ref type="formula" target="#formula_20">16</ref>) to zero, the update expression for Y is given as,</p><formula xml:id="formula_21">Y [𝑙+1] = [︁ (︀ W [𝑙] )︀ ⊤ W [𝑙] + 3𝛽I ]︁ −1 [︁ (︀ W [𝑙] )︀ ⊤ X [𝑙] + 𝛽 (︀ E [𝑙+1] + F [𝑙+1] + G [𝑙+1] )︀ − (︀ Φ [𝑙] 1 + Φ [𝑙] 2 + Φ [𝑙] 3 )︀ ]︁<label>(17)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.5.">Updating W:</head><p>The W sub-problem is given as follows.</p><formula xml:id="formula_22">W [𝑙+1] = argmin W 1 2 ‖X − WY‖ 2 F (18)</formula><p>Solution of the above equation can be found as,</p><formula xml:id="formula_23">W [𝑙+1] = [︁ X [𝑙] (︀ Y [𝑙+1] )︀ ⊤ ]︁[︁ Y [𝑙+1] (︀ Y [𝑙+1] )︀ ⊤ ]︁ −1<label>(19)</label></formula><p>Finally, the Lagrange multipliers are updated as follows:</p><formula xml:id="formula_24">Φ [𝑙+1] 1 = Φ [𝑙] 1 + 𝛽 (︀ Y [𝑙+1] − E [𝑙+1] )︀<label>(20)</label></formula><formula xml:id="formula_25">Φ [𝑙+1] 2 = Φ [𝑙] 2 + 𝛽 (︀ Y [𝑙+1] − F [𝑙+1] )︀<label>(21)</label></formula><formula xml:id="formula_26">Φ [𝑙+1] 3 = Φ [𝑙] 3 + 𝛽 (︀ Y [𝑙+1] − G [𝑙+1] )︀<label>(22)</label></formula><p>Convergence of the algorithm is ensured if, max</p><formula xml:id="formula_27">⎧ ⎨ ⎩ ⃦ ⃦ Y [𝑙+1] − E [𝑙] ⃦ ⃦ ∞ , ⃦ ⃦ E [𝑙+1] − E [𝑙] ⃦ ⃦ ∞ ⃦ ⃦ Y [𝑙+1] − F [𝑙] ⃦ ⃦ ∞ , ⃦ ⃦ F [𝑙+1] − F [𝑙] ⃦ ⃦ ∞ ⃦ ⃦ Y [𝑙+1] − G [𝑙] ⃦ ⃦ ∞ , ⃦ ⃦ G [𝑙+1] − G [𝑙] ⃦ ⃦ ∞ ⎫ ⎬ ⎭ &lt; 𝜖<label>(23)</label></formula><p>The overall process involved in the proposed time series activity clustering algorithm is summarized in Algorithm 1.</p><p>Once the representation matrix Y is obtained, an affinity matrix Q ∈ R 𝑘×𝑘 is calculated. The accuracy of clustering is highly dependent on the affinity matrix Q. A usual approach in obtaining the affinity matrix is as shown below <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref>.</p><formula xml:id="formula_28">Q = |Y| + |Y ⊤ | 2<label>(24)</label></formula><p>But, the graph so constructed do not take into account the intrinsic relationships of the withincluster data points. But for data containing temporal information, the within-cluster data points are highly correlated. In order to take advantage of this information, an affinity matrix Q is calculated as follows.</p><formula xml:id="formula_29">Q(𝑖, 𝑗) = 𝑦 ⊤ 𝑖 𝑦 𝑗 ‖𝑦 𝑖 ‖ 2 ‖𝑦 𝑗 ‖ 2<label>(25)</label></formula><p>where, ‖.‖ 2 represents the ℓ 2 norm operator. Update 𝛽 [𝑙+1] = 𝜂𝛽 [𝑙]   14:</p><formula xml:id="formula_30">𝑙 = 𝑙 + 1 15:</formula><p>Use <ref type="bibr" target="#b22">(23)</ref> to check the convergence 16: end while</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experimental Results and Analysis</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Dataset and Parameter Settings</head><p>To verify the performance of the proposed algorithm, it was tested on multiple human activity datasets. The datasets considered include the Gaming 3D (G3D) <ref type="bibr" target="#b21">[22]</ref>, Florence 3D (F3D) <ref type="bibr" target="#b22">[23]</ref>, UTKinect-Action 3D (UTK) <ref type="bibr" target="#b23">[24]</ref>, MSRC-Kinect12 (MSRC) <ref type="bibr" target="#b24">[25]</ref>, MSR Action 3D (MSRA) <ref type="bibr" target="#b25">[26]</ref>, and HDM14 <ref type="bibr" target="#b26">[27]</ref> datasets. By means of observation, the parameters 𝛾, 𝜆 1 , 𝜆 2 and 𝜂 are set to 5.2, 0.03, 18, and 0.7 respectively. The program development was done on a system with Intel Core i7 processor and a RAM of 16 GB, operating on 64-bit windows operating system with a clock frequency of 2.90 GHz.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Experimental Results</head><p>Experimental validation was done on the following methods.  Subspace clustering approaches based on self representation: In this method, clustering using state-of-the-art clustering approaches are performed on the generated affinity matrix. The clustering methods considered include Spectral Clustering (SC) <ref type="bibr" target="#b27">[28]</ref>, Orthogonal Matching Pursuit (OMP) <ref type="bibr" target="#b28">[29]</ref>, K-means (Km) <ref type="bibr" target="#b27">[28]</ref>, SSC <ref type="bibr" target="#b13">[14]</ref>, and Elastic Net Subspace Clustering (EnSC) <ref type="bibr" target="#b28">[29]</ref>. The results obtained using SSC is found to be superior to that of the counterparts <ref type="bibr" target="#b13">[14]</ref>.</p><p>SSC approaches with Data Pruning: In this method, the input skeletal data is pruned with strategies including min Φ <ref type="bibr" target="#b28">[29]</ref>, Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, Threshold Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, and Percentage Temporal SSC <ref type="bibr" target="#b28">[29]</ref>. Later, a feature matrix is generated from the pruned data sequences, followed by the application of SSC <ref type="bibr" target="#b13">[14]</ref> approach. This method converges quickly. Among others, the percentage temporal SSC approach gives better results.</p><p>TSC-Cov: This is a method that we had developed in one of our previous works. In this method, data re-framing was not performed and the kernel-based features were not incorporated in the covariance descriptor. But, we had used a new clustering approach named TSC-Cov for subspace clustering.</p><p>The performance of the proposed algorithm was evaluated against the above mentioned methods. The metrics used for quantitative evaluation include the accuracy, the Normalized Mutual Information (NMI), and the Adjusted Rand Index (ARI). The results obtained are tabulated in Tables (1-3).</p><p>For qualitative evaluation, the affinity matrix and clustering results obtained using the proposed method and the state-of-the-art methods are presented. Although the experiments were conducted on all the six datasets mentioned earlier, for the purpose of illustration, the results obtained with the UTK dataset <ref type="bibr" target="#b23">[24]</ref> is given in this paper.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the proposed time series activity clustering</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Affinity graphs of SSC, Percentage Temporal SSC, Threshold Temporal SSC, TSC-Cov, and Kernel and LRR imposed TSC (Proposed) approaches on UTK Dataset.</figDesc><graphic coords="9,121.96,572.43,348.86,65.86" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Clustering Results on UTK Dataset</figDesc><graphic coords="10,153.55,84.19,285.70,254.59" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>1 2 ‖X − WY‖ 2 F captures the reconstruction error, ||•|| 𝐹 represents the Frobenius norm operator, || • || * denotes the nuclear norm operator, and 𝜆 1 and 𝜆 2 are the balancing parameters.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Algorithm 1</head><label>1</label><figDesc>Time Series Activity Clustering Require: Skeletal data and parameters 𝜆 1 , 𝜆 2 , 𝜆 3 , 𝜂, 𝛾 and 𝛽</figDesc><table><row><cell cols="2">Ensure: Y ∈ R 𝑘×𝑘</cell><cell></cell></row><row><cell cols="3">1: Find Ψ using Eq. (1)</cell></row><row><cell cols="2">2: Find X using Ψ</cell><cell></cell></row><row><cell cols="3">3: Generate matrices Z, ̃︁ W, and L T</cell></row><row><cell cols="3">4: while 𝑛𝑜𝑡 𝑐𝑜𝑛𝑣𝑒𝑟𝑔𝑒𝑑 do</cell></row><row><cell>5:</cell><cell cols="2">Update E [𝑙+1] with Eq. (11)</cell></row><row><cell>6:</cell><cell cols="2">Update F [𝑙+1] with Eq. (13)</cell></row><row><cell>7:</cell><cell cols="2">Update G [𝑙+1] with Eq. (15)</cell></row><row><cell>8:</cell><cell cols="2">Update Y [𝑙+1] with Eq. (17)</cell></row><row><cell>9:</cell><cell cols="2">Update W [𝑙+1] with Eq. (19)</cell></row><row><cell>10: 11: 12:</cell><cell>Update Φ [𝑙+1] 1 Update Φ [𝑙+1] 2 Update Φ [𝑙+1] 3</cell><cell>with Eq. (20) with Eq. (21) with Eq. (22)</cell></row><row><cell>13:</cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Comparison of Clustering accuracy (%)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Dataset</head><p>SSC <ref type="bibr" target="#b13">[14]</ref> Percentage Temporal SSC <ref type="bibr" target="#b28">[29]</ref> TSC-Cov TSAC (Proposed) G3D <ref type="bibr" target="#b21">[22]</ref> 65.16 66.04 90.04 92.65 F3D <ref type="bibr" target="#b22">[23]</ref> 61.86 63.72 81.39 81.58 UTK <ref type="bibr" target="#b23">[24]</ref> 74.37 72.36 89.44 90.01 MSRC <ref type="bibr" target="#b24">[25]</ref> 73.42 80.60 81.09 82.00 MSRA <ref type="bibr" target="#b25">[26]</ref> 58 TSC-Cov TSAC (Proposed) G3D <ref type="bibr" target="#b21">[22]</ref> 0.719 0.708 0.953 0.961 F3D <ref type="bibr" target="#b22">[23]</ref> 0.716 0.709 0.872 0.875 UTK <ref type="bibr" target="#b23">[24]</ref> 0.709 0.672 0.899 0.911 MSRC <ref type="bibr" target="#b24">[25]</ref> 0.720 0.762 0.887 0.890 MSRA <ref type="bibr" target="#b25">[26]</ref> 0.700 0.720 0.958 0.972 HDM14 <ref type="bibr" target="#b26">[27]</ref> 0.754 0.771 0.893 0.901</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Comparison of ARI Dataset SSC <ref type="bibr" target="#b13">[14]</ref> Percentage Temporal SSC <ref type="bibr" target="#b28">[29]</ref> TSC-Cov TSAC (Proposed) G3D <ref type="bibr" target="#b21">[22]</ref> 0.499 0.479 0.847 0.851 F3D <ref type="bibr" target="#b22">[23]</ref> 0.548 0.539 0.781 0.787 UTK <ref type="bibr" target="#b23">[24]</ref> 0.547 0.499 0.804 0.819 MSRC <ref type="bibr" target="#b24">[25]</ref> 0.551 0.714 0.773 0.781 MSRA <ref type="bibr" target="#b25">[26]</ref> 0.435 0.456 0.881 0.895 HDM14 <ref type="bibr" target="#b26">[27]</ref> 0.439 0.484 0.753 0.772 Fig. <ref type="figure">2</ref> visualizes the affinity matrices generated using SSC <ref type="bibr" target="#b13">[14]</ref>, Percentage Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, Threshold Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, TSC-Cov, and the proposed method while working on the UTK dataset. We can observe that among others, the affinity matrix generated using the proposed method have much denser block diagonal structure. This is an indication of quality of the clustering process.</p><p>Fig. <ref type="figure">3</ref> shows the clustering results obtained on the UTK dataset. For visual analysis, each cluster is assigned a unique color. The figure shows a comparison between SSC <ref type="bibr" target="#b13">[14]</ref>, Percentage Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, Threshold Temporal SSC <ref type="bibr" target="#b28">[29]</ref>, TSC-Cov, and the proposed method with reference to the true labels. From Fig. <ref type="figure">3</ref> we can observe that clustering results obtained with the proposed method are comparatively better than the other methods.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>The paper proposes a new method for clustering of human activity sequences in an efficient way. The proposed method involves the extraction of features from raw input data with the help of a kernel-based robust covariance descriptor. The optimization model developed uses the combined advantage of LRR and LSR based subspace clustering approaches. The concept of temporal Laplacian regularized dictionary learning is introduced in order to learn an effective representation matrix from the extracted data features. With the help of ADMM approach, the solution for the optimization problem is obtained. Performance of the proposed approach is compared with that of the state-of-the-art approaches in terms of accuracy, NMI, and ARI. Experimental results validate superiority of the proposed method in obtaining better clustering results as compared to that of the counterparts. Motion capture data often suffers from corruptions in the recorded information. To address this problem, robust human activity clustering algorithms can be developed in the future. Also, mocap information can be combined with other modalities of human activity data to achieve improved clustering results in challenging scenarios.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A survey on video-based human action recognition: recent updates, datasets, challenges, and applications</title>
		<author>
			<persName><forename type="first">P</forename><surname>Pareek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Thakkar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="2259" to="2322" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A comprehensive survey of vision-based human action recognition methods</title>
		<author>
			<persName><forename type="first">H.-B</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-X</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D.-S</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page">1005</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Human activity recognition: A survey</title>
		<author>
			<persName><forename type="first">C</forename><surname>Jobanputra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bavishi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Doshi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procedia Computer Science</title>
		<imprint>
			<biblScope unit="volume">155</biblScope>
			<biblScope unit="page" from="698" to="703" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Human action recognition and prediction: A survey</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Vision</title>
		<imprint>
			<biblScope unit="volume">130</biblScope>
			<biblScope unit="page" from="1366" to="1401" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Sensor-based and visionbased human activity recognition: A comprehensive survey</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">M</forename><surname>Dang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Min</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Piran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Moon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">108</biblScope>
			<biblScope unit="page">107561</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Skeleton-based human activity recognition using convlstm and guided feature learning</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tiwari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">M</forename><surname>Pandey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Akbar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Soft Computing</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="877" to="890" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Ongoing human action recognition with motion capture</title>
		<author>
			<persName><forename type="first">M</forename><surname>Barnachon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bouakaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Boufama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Guillou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">47</biblScope>
			<biblScope unit="page" from="238" to="247" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A depth camera-based human activity recognition via deep learning recurrent neural network for health and social care services</title>
		<author>
			<persName><forename type="first">S</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Al-Masni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Al-Antari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Z</forename><surname>Uddin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-S</forename><surname>Kim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procedia Computer Science</title>
		<imprint>
			<biblScope unit="volume">100</biblScope>
			<biblScope unit="page" from="78" to="84" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Human activity recognition from 3d data: A review</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="70" to="80" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Symbiotic graph neural networks for 3d skeleton-based human action recognition and motion prediction</title>
		<author>
			<persName><forename type="first">M</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Tian</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Tran. on PAMI</title>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bagnall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lines</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bostrom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Large</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Keogh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Mining and Knowledge Discovery</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="606" to="660" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Subspace clustering for high dimensional data: a review</title>
		<author>
			<persName><forename type="first">L</forename><surname>Parsons</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Haque</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Acm sigkdd explorations newsletter</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="90" to="105" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Low rank subspace clustering (lrsc)</title>
		<author>
			<persName><forename type="first">R</forename><surname>Vidal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Favaro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="47" to="61" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Sparse subspace clustering: Algorithm, theory, and applications</title>
		<author>
			<persName><forename type="first">E</forename><surname>Elhamifar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Vidal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Tran. on PAMI</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="2765" to="2781" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Spatial-spectral structured sparse low-rank representation for hyperspectral image super-resolution</title>
		<author>
			<persName><forename type="first">J</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Liao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>.-W. Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Philips</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Tran. on Image Processing</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="3084" to="3097" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">A joint sparse and correlation induced subspace clustering method for segmentation of natural images</title>
		<author>
			<persName><forename type="first">J</forename><surname>Francis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Madathil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">N</forename><surname>George</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 17th India Council Int. Conf. (INDICON)</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Robust spectral subspace clustering based on least square regression</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Processing Letters</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="1359" to="1372" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Kernelized covariance for action recognition</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cavazza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zunino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Biagio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Murino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2016 23rd Int. Conf. on Pattern Recognition (ICPR)</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="408" to="413" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Low-rank matrix approximation with manifold regularization</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Tran. on PAMI</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="1717" to="1729" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">𝑝-laplacian regularization for scene recognition</title>
		<author>
			<persName><forename type="first">W</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cheng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Tran. on Cybernetics</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="page" from="2927" to="2940" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Part-based data analysis with masked non-negative matrix factorization</title>
		<author>
			<persName><forename type="first">G</forename><surname>Casalino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">D</forename><surname>Buono</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mencar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Int. Conf. on Computational Science and Its Applications</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="440" to="454" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Hierarchical transfer learning for online recognition of compound actions</title>
		<author>
			<persName><forename type="first">V</forename><surname>Bloom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Argyriou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Makris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Vision and Image Understanding</title>
		<imprint>
			<biblScope unit="volume">144</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses</title>
		<author>
			<persName><forename type="first">L</forename><surname>Seidenari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Varano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Berretti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Del Bimbo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pala</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2013 IEEE Conf. on CVPR -Workshops</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="479" to="485" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">View invariant human action recognition using histograms of 3d joints</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Aggarwal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Computer Society Conf.on CVPR -Workshops</title>
				<imprint>
			<date type="published" when="2012">2012. 2012</date>
			<biblScope unit="page" from="20" to="27" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Instructing people for training gestural interactive systems</title>
		<author>
			<persName><forename type="first">S</forename><surname>Fothergill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">M M</forename></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kohli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nowozin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CHI &apos;12 Proc. of the SIGCHI Conf. on Human Factors in Computing Systems</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="1737" to="1746" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Action recognition based on a bag of 3d points</title>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Computer Society Conf. on CVPR -Workshops</title>
				<imprint>
			<date type="published" when="2010">2010. 2010</date>
			<biblScope unit="page" from="9" to="14" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Documentation Mocap Database HDM05</title>
		<author>
			<persName><forename type="first">M</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Röder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Clausen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Eberhardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Krüger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Weber</surname></persName>
		</author>
		<idno>CG-2007-2</idno>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>Universität Bonn</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Minimum entropy, k-means, spectral clustering</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Choi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Int. Joint Conf. on Neural Networks (IEEE Cat. No.04CH37541)</title>
				<imprint>
			<date type="published" when="2004">2004. 2004</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="117" to="122" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Subspace clustering for action recognition with covariance representations and temporal pruning</title>
		<author>
			<persName><forename type="first">G</forename><surname>Paoletti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cavazza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Beyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Del Bue</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">25th Int. Conf. on Pattern Recognition (ICPR)</title>
				<imprint>
			<date type="published" when="2020">2020. 2021</date>
			<biblScope unit="page" from="6035" to="6042" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
