<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards User Recognition by Shallow Web Traffic Inspection</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Marino</forename><surname>Miculan</surname></persName>
							<email>marino.miculan@uniud.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics, Computer Science and Physics</orgName>
								<orgName type="institution">University of Udine</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gian</forename><forename type="middle">Luca</forename><surname>Foresti</surname></persName>
							<email>gianluca.foresti@uniud.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics, Computer Science and Physics</orgName>
								<orgName type="institution">University of Udine</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Claudio</forename><surname>Piciarelli</surname></persName>
							<email>claudio.piciarelli@uniud.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics, Computer Science and Physics</orgName>
								<orgName type="institution">University of Udine</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Towards User Recognition by Shallow Web Traffic Inspection</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B09AEBB399D3762C48EEC20DCC3221B7</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We consider the problem of web user recognition, or web traffic de-anonymization: given a set of users, is it possible to identify which user has generated a given web traffic with only a shallow packet inspection (that is, without looking inside the packet payloads)?</p><p>We propose to address this problem by means of machine learning (ML) techniques, in particular clustering and supervised classification. The basic idea is that each user can be identified by their browsing habits, and these habits can be described by a suitable set of features: click frequency, permanence time on web pages, amount of downloaded data, etc. In this paper we introduce these features, and show how these can be derived from the data obtained only from packet headers and their arrival time. Finally, we show the effectiveness of this approach with some preliminary tests and experiments.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Nowadays, it is important to be able to identify the users accessing the Internet both for commercial and forensics purposes. On one side, Internet Service Providers and Content Providers can take advantage of this analysis, in order to personalize and improve their services; on the other, it is important to identify users in order to ascertain the responsibility for harmful or even criminal actions.</p><p>In fact, ISPs are already required to keep access log files, whose acquisition from the authorities is regulated by Directives such as 2006-24-CE, both for cybercrimes against persons (such as online fraud and identity theft) and for attacks against a server or a network. However, these access logs record only limited information, such as connection times, transfer amounts, and visited web site addresses, which in general are not sufficient to trace the specific author of an offense, but only the holder of the network connection contract. This is the case of networks which access the Internet via one or few public IP addresses by means of the well-known network address translation technique, such as in the case of home networks, small enterprises, commercial premises offering WiFi connection to their customers, etc.. Another scenario is that of workstations accessible by many users in public or semi-public environments, like college laboratories, hotel lobbies, internet cafés, etc.. In these cases, we cannot associate the traffic from a given IP address to a specific user, because a single public IP address is shared among several users, possibly at the same time: family members and their visiting friends, employees and their collaborators, students, hotel guests, customers, etc.. One could argue that more useful information can be obtained by looking for relevant data (e.g. usernames, email addresses) inside the payloads of IP packets. This technique, called deep packet inspection (DPI), is well-known and can be very effective <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b7">8]</ref>, but it can be applied only if the traffic is not encrypted. Nowadays most web traffic (especially that carrying identification data) is encrypted at the transport level by means of SSL and TLS protocols; actually an increasing number of websites is adopting encryption protocols, and therefore DPI will be less and less applicable. Moreover, DPI raises important privacy issues, because it allows the inspector to access the whole traffic content, not only the data needed for user identification <ref type="bibr" target="#b1">[2]</ref>.</p><p>Therefore, the problem is: given a set of users, is it possible to identify which user has generated a given web traffic, by means of shallow packet inspection? By "shallow" we mean that the only web traffic data we are allowed to consider are those an ISP can normally access for providing its service: the network (IP) header, the transport (TCP/UDP) header, the sizes and frequency of packets, etc., but not the content of the TCP payload. Shallow packet inspection is a novel technique that only very recently has been investigated by the research community, not for de-anonymization but for traffic classification <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>In this paper, we propose to address this problem by means of machine learning (ML) techniques, in particular clustering and supervised classification. The basic idea is that each user can be identified by their browsing habits, and that these habits can be described by a suitable set of features, such as traffic size, click frequency, permanence time on web pages, hour of the day of the activity, etc. In this paper we introduce some of these features, and show how these can be derived from the data obtained by shallow packet inspection. To this end, we first introduce the notion of click as the basic event to consider in the analysis. A "click" represents the voluntary action performed by the user when clicking on a hyperlink/button on a web page. Therefore, a click subsumes the whole traffic (that is, all the HTTP(S) requests and replies) caused by this action. Click data are obtained by partitioning the packet flow to/from the observed IP address, using clustering techniques. Then, on the flow of clicks we identify suitable features for the classification. Some features are intuitive (e.g. counterpart IP address, traffic size), but others are less obvious, such as the time of the day when the click is performed, or the "dwell time" on a web page. These data are related to the browsing habits of each user, and hence can be used as the basis for the supervised classification algorithms. We consider different algorithms; each of them is trained and tested with the same sets of click flows. The results are encouraging: some algorithms yield high classification precision.</p><p>The rest of the paper is organized as follows. In Section 2 we present in detail the problem under consideration. The solution we propose is described in Section 3, and experimental results are reported in Section 4. Finally in Section 5 we draw some conclusions and outline future work. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Problem description and analysis</head><p>Let us consider the scenario shown in Figure <ref type="figure" target="#fig_0">1</ref>. In this scenario, there is a private network where several users u 1 , . . . , u n can browse the Web using the same client computer. During a session, the client computer is used by only one user. The browsing activity of the user generates a sequence of HTTP(S) requests towards various WWW servers on the Internet, and corresponding replies. In turn, these requests yield a flow of IP packets from the client to the web servers and back. These flows go through the border router and the ISP router(s), where can be examined by an observer. Very often, the private network uses a non-routable address space (such as 10.0.0.0/8 or 192.168.0.0/16). In this case the border router performs a NAT/PAT translation <ref type="bibr" target="#b9">[10]</ref>: in each outgoing packet the source addresses and port are replaced with the public IP address assigned to the border router by the ISP-and dually for the incoming packets. The net effect is that, from the public network viewpoint, the whole private network appears as a single host with the public IP address.</p><p>Using a packet analyzer, the observer can gather the following data for each packet:</p><p>-Arrival time at the router; -Total length of the packet; -Source IP address and port number: for packets coming from the border router, the source IP address is the public IP assigned to the local router by the ISP, and the port is a dynamic port on such router. For packets going to the border router, the source IP address and port number are those of the contacted web server; -Destination IP address and port number: for packets coming from the local router, the destination IP address and port number are those of the contacted web server; for packets going to the local router, the destination IP address is the public IP assigned to the local router by the ISP, and the port is a dynamic port on such router.</p><p>Other data in the headers can be ignored because either not relevant (such as TOS, or fragmentation details), or non informative (the version of IP is almost always 4, the trasport level protocol is always TCP, etc.). It is important to notice that these data cannot be encrypted, otherwise the routers would not be able to route the packet. Following the shallow packet inspection policy, we do not analyze the content of TCP segments; very likely these segments carry SSL/TLS encrypted payloads, which cannot be analyzed further. Therefore, for each web session we can obtain a file, called web session log, of tuples of the following form: arrival time, packet length, source IP, source port, destination IP, destination port .</p><p>Since we intend to apply supervised classification algorithms, we assume to be able to obtain a suitable training set in order to "learn" the browsing habits of each user. A training set is a set T S = { L 1 , u i1 , . . . , L k , u i k } of web session logs associated to the corresponding user. Here, i k is the index of the user generating log L k . This set can be build by observing the traffic when we know the actual identity of the user browsing at that moment.</p><p>Then, the classification problem can be formulated as follows:</p><p>given a training set T S for users u 1 , . . . u n and a web session log L generated by one of these users, is it possible to determine which user has generated L?</p><p>The criteria for evaluating a classifier are the usual ones:</p><p>Accuracy: the percentage of web session logs correctly classified; Recall: the percentage of web session logs correctly assigned to a user, with respect to all logs generated by that user; Precision: the percentage of web session logs correctly assigned to a user, with respect to all logs assigned to that user; F-measure: the harmonic average of recall and precision.</p><p>In the next section we propose a machine learning-based solution to this problem.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Solution proposal</head><p>In order to recognize users by means of shallow packet inspection, we propose the architecture depicted in Figure <ref type="figure" target="#fig_1">2</ref>. When building the training set, we assume that users can be identified by the IP address of their workstation, thus no user can access to more than one workstation, and no workstation can be used by more than one user. Moreover, we assume that no NAT policy is implemented. These requirements are only needed to support the training phase of a supervised classifier, where each data sample must be labeled with the correct classification; user identification is not needed in classification phase, except for performance measurement. The whole network traffic is logged by a sniffer and subsequently filtered and pre-processed to collect only the data relevant for the system. Pre-processing also includes a clustering step, in which data associated to the same user action are grouped to identify meaningful high-level data that are fed to the classifier for training or evaluation. In order to preserve user privacy, pre-processing also replaces source IP addresses with unique identifiers (User1, User2, etc.). Hence, despite the system internally stores the address/identifier associations (which are needed to guarantee a coherent labeling through time), the final data are pseudonymized and can be safely shared.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Feature extraction and filtering</head><p>The sniffer acquires and logs all the network traffic coming to/from the network; it is thus placed either in the local network itself (where it can access the data by enabling the promiscuous mode of the network card) or in a bottleneck device such a network router. In case of small networks, it can be implemented using publicly available software, such as WireShark. However, in most cases a full log of network traffic quickly leads to extremely large log files, which are both impractical to store and pose privacy issues, since they may contain data outside the scope of the proposed system. We thus adopt filtering rules on the sniffer, in order to log only relevant traffic. In particular, we assume that most user actions generate TCP/IP traffic, thus any other packet type is silently ignored. This include both user-generated data, such as UDP/IP packets, as well as network management data, such as ARP, SNMP, RIP packets etc.. Moreover, in this work we choose to focus only on a specific type of data, namely web traffic, as the Web is a popular service that may contain several user-distinctive features to ease the classification task. The sniffer is thus configured to acquire only TCP traffic to or from port 80 (HTTP) or 443 (HTTPS). The encryption of HTTPS connections is not an issue, since the proposed system performs a shallow inspection, without analyzing the packet payload. Finally, to further reduce the amount of data to be stored, we extract only relevant features form each packet. As mentioned in Section 2, the final collected raw data are:</p><p>packet arrival time; client IP address and TCP port; server IP address and TCP port; packet length.</p><p>The client IP address is then pseudonymized as described at the beginning of this section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Feature pre-processing</head><p>Since the goal of the proposed system is to identify users by their web navigation behaviors, it is important that data can be associated to voluntary user actions. This task is not trivial since network traffic generated by modern web navigation can be loosely connected to user actions. Let us consider for example the basic action of clicking on a hyperlink. The corresponding network traffic delivers the required web page, but also new, parallel connections deliver other page contents (e.g. images, cookies, etc.). Moreover, new connections can be established to other web servers, such as advertisement-delivery services, profiling services etc. Thus, in order to work on a higher abstraction level we introduce the notion of click, defined as the set of all network traffic generated by a user action, such as clicking on a hyperlink. Clicks can be extracted by temporally clustering the data packets: groups of data packets with close arrival times are considered part of the same click, even if originating from different servers.</p><p>Since the number of expected clicks is unknown a priori, no algorithms requiring an initial knowledge on the number of clusters is suitable, thus excluding popular clustering techniques such as k-means or Gaussian Mixture Models. Moreover, we require hard clustering (cluster membership is a binary choice) and explicit outlier modeling, since not all the network traffic could belong to a click. These considerations motivate the choice of DBSCAN <ref type="bibr" target="#b2">[3]</ref> as clustering algorithm. DBSCAN uses a density-based approach, where clusters are defined as groups of high-density samples. Formally, given a set of samples P , we give the following definitions:</p><p>p ∈ P is a core point if at least m points q 1 . . . q m ∈ P exist such that</p><formula xml:id="formula_0">P − q i ≤ ∀i ∈ [1 . . . m],</formula><p>where m, are the algorithm parameters; q ∈ P is directly-reachable from p ∈ P if p − q ≤ and p is a core point; q ∈ P is density-reachable from p ∈ P if there exists a path p 1 . . . p n such that p 1 = p, p n = q and p i+1 is directly-reachable from</p><formula xml:id="formula_1">p i ∀i ∈ [1 . . . n − 1].</formula><p>Given a core point p, its cluster is then defined as the set of all the points that are density-reachable from p. Points that are not density-reachable by any other point are marked as outliers. Figure <ref type="figure" target="#fig_2">3</ref> shows the effect of DBSCAN applied to the arrival time of a set of packets in order to identify clicks and outliers.</p><p>For each click, we compute the following features: user ID; main site, i.e. the IP address of the destination server of the first packet in the click; timestamp, defined as the arrival time of the first packet in the click; inter-click time, defined as the time passed since the last click from the same user to the same main site (set to 0 if it is the first one); total amount of data, defined as the sum of all packet lengths in the click; total number of secondary sites, i.e. destination servers different from the main site.</p><p>Furthermore, by analyzing all the clicks originated from the same user, we define a session log as a set of statistics about the acquired clicks. A session log is thus a data sample containing:</p><p>user ID; main site; session start time, defined as the timestamp of the first click; session end time, defined as the timestamp of the last click; session duration, defined as the the difference between session end time and start time; total number of clicks in the session; average inter-click time, defined as session duration / total number of clicks; average click data length, defined as total amount of data / total number of clicks; average number of secondary sites.</p><p>As a final pre-processing step, all the numerical data are standardized, since this is required by many machine learning algorithms. Standardization is achieved by mean removal and variance scaling: given a set of feature values {f 1 . . . f n }, each feature value f i is replaced by its standardized version fi defined as:</p><formula xml:id="formula_2">fi = f i − f 1/n n j=1 (f j − f ) 2<label>(1)</label></formula><p>where f = 1/n n j=1 f j . Feature means and standard deviations are computed on the training set only. Test sets are standardized using the same values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Classification algorithms</head><p>The acquired session data are used to train a machine learning classifier. As features, we consider all the numerical data of a session. The main site, despite being a strong hint for user identification, is currently discarded since its categorical, rather than numerical, nature poses extra processing difficulties that will be addressed in a future work. The User IDs of each session are used as sample labels.</p><p>In order to classify the data, we considered the following algorithms:</p><p>Naive Bayes Naive Bayes classifiers <ref type="bibr" target="#b12">[13]</ref> define P (y|x 1 . . . x n ) as the probability of class y given the features x 1 . . . x n . Under the naive (hence the name) assumption of conditional independence between every pair of features given the value of the class variable, it can be proven that:</p><formula xml:id="formula_3">P (y|x 1 . . . x n ) ∝ P (y) n i=1 P (x i |y)<label>(2)</label></formula><p>and the class estimate ŷ is thus defined as: ŷ = arg max y P (y)</p><formula xml:id="formula_4">n y=1 P (x i |y)<label>(3)</label></formula><p>where P (y) and P (x i |y) can be estimated from data using maximum a posteriori (MAP) estimation. In this work we adopted a Gaussian Naive Bayes model, where P (x i |y) is assumed to be a Gaussian function.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>K-Star</head><p>The K-Star classifier <ref type="bibr" target="#b6">[7]</ref>, or K * , is an instance-based classifier, meaning that the class of an instance is based upon the class of those training instances similar to it, as determined by some similarity function. Specifically, K-Star adopts a entropy-based distance function, calculated by mean of the complexity of transforming an instance into another.</p><p>Support Vector Machines Linear Support Vector Machines <ref type="bibr" target="#b0">[1]</ref> are based on the idea that an optimal linear classifier maximizes the margin, this is the width of the strip parallel to the classification hyperplane that separates the two classes.</p><p>The solution is found by solving a constrained optimization model leading to the following classification function:</p><formula xml:id="formula_5">f (x) = sgn n i=1 y i α i (x • x i ) + b<label>(4)</label></formula><p>where x i is a sample from the training set and y i is the corresponding class label, while α and b are found by solving the optimization problem. The solution is actually sparse since α i = 0 for most of the data, except the few ones lying on the margin (support vectors). SVMs became popular because they can be easily extended to the non-linear case by means of kernel methods.</p><p>C4.5 The C4.5 algorithm <ref type="bibr" target="#b8">[9]</ref> is a decision tree building algorithm based on its precursor ID3. The decision tree is built using the concept of information entropy: at each node, C4.5 chooses the feature that most effectively splits the data associated to the branch. The effectiveness of the split is measured in terms of information gain (difference in entropy) of the feature. Once built, the decision tree can be easily used to classify new data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Tests and evaluation</head><p>In oder to evaluate the system performances, we tested the system on a Local Area Network where 10 users (both male and female, in the age range of 20-45 years) were asked to visit web pages as in their normal daily routine. We did not consider the case of a malicious user deliberately trying to hide their network traffic pattern. Users were informed about privacy issues, to clarify that only web traffic metadata was logged and no deep inspection was performed. This way, users felt free to use the web as in their normal activity. Data were acquired along 10 sessions, each one 10 minutes long. On average, we collected ∼ 300 clicks per user. The relatively small amount of data motivates the choice of the algorithms presented in Section 3, since more sophisticated techniques such as deep neural networks would require lager datasets. After pre-processing, the dataset was split in a training and a test set with different ratios to evaluate the performances on low amounts of training data. Tests were performed using respectively 20%, 50% and 80% of the original data as training set. Results obtained with the four classifiers are shown in Table <ref type="table" target="#tab_0">1</ref>.</p><p>As it can be seen, all the classifiers achieved good performances except Support Vector Machines. This could be explained by a poor choice of training parameters, since SVM results heavily depends on the choice of kernel and kernel parameters. As a future work, we plan to further investigate this aspect. Among the remaninig classifiers, C4.5 performed best, reaching high accuracy levels even with a small training set (20% of total data). The preliminary results are thus encouraging, and in the next future we aim to test the system with a larger user base. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions</head><p>The ability to identify web users by analyzing their network traffic can have multiple applications, from user profiling to digital forensics. In this paper we investigated the possibility of identifying users only by means of shallow inspection of HTTP(S) network traffic. Shallow inspection, in which the content of the packet payload is not analyzed, is motivated both by privacy issues and by technological factors: nowadays, the increasing adoption of encrypted connections is making deep inspection mostly useless. Despite the few amount of data gathered by shallow inspection, we proposed a data pre-processing method to extract high-level features that could be relevant for user identification, such as inter-click time intervals, time spent on a single web page, etc.. We tested four different classifiers on a small dataset obtaining encouraging preliminary results.</p><p>As a future work, we plan to acquire a larger dataset in order to test more complex classifiers such as deep neural networks. Moreover, we intend to investigate the reasons for the relatively low performance of the SVM classifier, in particular concerning the choice of the kernel and kernel parameters. We also plan to evaluate if automatic deep feature extraction techniques can actually outperform our manually-defined high-level feature set <ref type="bibr" target="#b3">[4]</ref>. Finally, we will also focus on proper representation and processing of categorical data, in order to handle non-numerical features such as server IP addresses.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Example scenario.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Architecture of the web session classifier.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Histogram of the number of detected packets while clicking on three links. Colored areas under the histogram show the three clicks detected by the proposed clustering procedure.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Classification results.</figDesc><table><row><cell>Classifier</cell><cell cols="4">Accuracy (%) Precision (%) Recall (%) F-Score (%)</cell></row><row><cell>Naive Bayes 20%</cell><cell>95</cell><cell>96</cell><cell>95</cell><cell>95</cell></row><row><cell>Naive Bayes 50%</cell><cell>97</cell><cell>98</cell><cell>97</cell><cell>97</cell></row><row><cell>Naive Bayes 80%</cell><cell>96</cell><cell>96</cell><cell>96</cell><cell>96</cell></row><row><cell>K-Star 20%</cell><cell>75</cell><cell>76</cell><cell>75</cell><cell>75</cell></row><row><cell>K-Star 50%</cell><cell>84</cell><cell>84</cell><cell>84</cell><cell>84</cell></row><row><cell>K-Star 80%</cell><cell>84</cell><cell>86</cell><cell>84</cell><cell>85</cell></row><row><cell>SVM 20%</cell><cell>44</cell><cell>69</cell><cell>44</cell><cell>45</cell></row><row><cell>SVM 50%</cell><cell>54</cell><cell>69</cell><cell>53</cell><cell>55</cell></row><row><cell>SVM 80%</cell><cell>59</cell><cell>75</cell><cell>59</cell><cell>62</cell></row><row><cell>C4.5 20%</cell><cell>97</cell><cell>97</cell><cell>97</cell><cell>97</cell></row><row><cell>C4.5 50%</cell><cell>99</cell><cell>99</cell><cell>99</cell><cell>99</cell></row><row><cell>C4.5 80%</cell><cell>99</cell><cell>100</cell><cell>100</cell><cell>100</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgments We thank Clelia Bincoletto for preliminary work and experiments on the subject of this paper.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Partially supported by UniUd PRID 2017 ENCASE and by Italy-Singapore bilateral technology cooperation project PRESNET.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">An introduction to support vector machines and other kernel-based learning methods</title>
		<author>
			<persName><forename type="first">N</forename><surname>Cristianini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shawe-Taylor</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>Cambridge university press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The legality of deep packet inspection</title>
		<author>
			<persName><forename type="first">A</forename><surname>Daly</surname></persName>
		</author>
		<idno type="DOI">10.2139/ssrn.1628024</idno>
		<ptr target="https://doi.org/10.2139/ssrn.1628024" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Communications Law &amp; Policy</title>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A density-based algorithm for discovering clusters in large spatial databases with noise</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Kriegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD-96)</title>
				<meeting>the Second International Conference on Knowledge Discovery and Data Mining (KDD-96)</meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="volume">96</biblScope>
			<biblScope unit="page" from="226" to="231" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Deep learning</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bach</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>MIT press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Advanced algorithms for fast and scalable deep packet inspection</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Turner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Williams</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Architecture for Networking and Communications systems</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2006">2006. 2006</date>
			<biblScope unit="page" from="81" to="92" />
		</imprint>
	</monogr>
	<note>ACM/IEEE Symposium on</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Deep packet: A novel approach for encrypted traffic classification using deep learning</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lotfollahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Shirali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Siavoshani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Saberian</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1709.02656</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Tool condition monitoring using k-star algorithm</title>
		<author>
			<persName><forename type="first">S</forename><surname>Painuli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Elangovan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sugumaran</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="2638" to="2643" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Parsons</surname></persName>
		</author>
		<title level="m">Deep Packet Inspection in Perspective: Tracing its lineage and surveillance potentials</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
		<respStmt>
			<orgName>Surveillance Studies Centre, Queen&apos;s University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">C4. 5: programs for machine learning</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Quinlan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<publisher>Elsevier</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">IP Network Address Translator (NAT) Terminology and Considerations</title>
		<author>
			<persName><forename type="first">P</forename><surname>Srisuresh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Holdrege</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>The Internet Society</publisher>
			<biblScope unit="volume">2663</biblScope>
			<pubPlace>rFC</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Feature extraction and visualization for network pcapng traces</title>
		<author>
			<persName><forename type="first">R</forename><surname>Velea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ciobanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Gurzau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">V</forename><surname>Patriciu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Control Systems and Computer Science (CSCS)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="311" to="316" />
		</imprint>
	</monogr>
	<note>21st International Conference on</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Network traffic anomaly detection using shallow packet inspection and parallel k-means data clustering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Velea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ciobanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Mȃrgȃrit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Bica</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Studies in Informatics and Control</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="387" to="396" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">The optimality of naive Bayes</title>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Seventeenth International Florida Artificial Intelligence Research Society Conference</title>
				<meeting>Seventeenth International Florida Artificial Intelligence Research Society Conference</meeting>
		<imprint>
			<date type="published" when="2004">2004. 2004</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
