<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Non-invasive AI-powered Diagnostics: The case of Voice-Disorder Detection -Vision paper</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Gabriele</forename><surname>Ciravegna</surname></persName>
							<email>gabriele.ciravegna@polito.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Politecnico di Torino</orgName>
								<address>
									<addrLine>Corso Duca degli Abruzzi</addrLine>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alkis</forename><surname>Koudounas</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Politecnico di Torino</orgName>
								<address>
									<addrLine>Corso Duca degli Abruzzi</addrLine>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marco</forename><surname>Fantini</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">ENT Unit</orgName>
								<orgName type="institution">San Feliciano Hospital</orgName>
								<address>
									<settlement>Rome</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tania</forename><surname>Cerquitelli</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Politecnico di Torino</orgName>
								<address>
									<addrLine>Corso Duca degli Abruzzi</addrLine>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Elena</forename><surname>Baralis</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Politecnico di Torino</orgName>
								<address>
									<addrLine>Corso Duca degli Abruzzi</addrLine>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Erika</forename><surname>Crosetti</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">ENT Clinic -Head and Neck Cancer Unit</orgName>
								<orgName type="institution">San Giovanni Bosco Hospital</orgName>
								<address>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giovanni</forename><surname>Succo</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">ENT Clinic -Head and Neck Cancer Unit</orgName>
								<orgName type="institution">San Giovanni Bosco Hospital</orgName>
								<address>
									<settlement>Turin</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<settlement>Paestum</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Non-invasive AI-powered Diagnostics: The case of Voice-Disorder Detection -Vision paper</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">9B1136C8295B21EC8074D796D12EBA14</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:17+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Artificial Intelligence</term>
					<term>Non-invasive diagnostics</term>
					<term>Voice disorder recognition</term>
					<term>Voice analysis</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper proposes a novel pipeline for non-invasive diagnosis and monitoring in healthcare, leveraging artificial intelligence (AI). The pipeline allows individuals to record various health data using everyday devices and analyze it via AI algorithms on a cloud-based platform. Experimental results on voice disorder detection demonstrate the effectiveness of the proposed approach when compared to existing solutions. Additionally, we discuss the positive impact of the pipeline on diagnosis, prognosis, and monitoring, emphasizing its non-invasive nature. Overall, we think the proposed pipeline might contribute to advancing AI-driven healthcare solutions with implications for global healthcare delivery.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Artificial Intelligence (AI) is increasingly integrated into healthcare, offering opportunities to improve diagnostics, treatment, and patient care. Through machine learning algorithms, AI systems can analyze medical data, providing clinical decision support and personalized treatment options <ref type="bibr" target="#b0">[1]</ref>. Non-invasive diagnostics is a critical aspect of modern healthcare delivery, offering patients a less intrusive and more comfortable care experience. This holds for both the diagnostic and the monitoring processes, improving the overall quality of life for patients <ref type="bibr" target="#b1">[2]</ref>. Beyond enhancing patient comfort, these methods also improve accessibility to healthcare services, particularly for underserved populations or those with limited access to specialized medical facilities.</p><p>We propose an AI-based framework for non-invasive diagnostics that integrates various data modalities, including voice recordings, self-pictures, typing patterns, ECG readings, and sleep analysis, among others. These diverse sources of data are collected through everyday devices such as computers, smartphones, and smartwatches, enabling convenient and continuous monitoring of individual health metrics. Subsequently, these data are fed into cloud-based applications where advanced AI algorithms analyze them, identifying patterns that may be indicative of underlying health issues. Also, the integration of multiple data sources allows for a holistic understanding of an individual's health, facilitating early detection and personalized intervention strategies.</p><p>The envisioned framework has the potential to improve healthcare delivery as well as have economic and technological impacts. First, by enabling large scale screening, it may increase early detection, allowing for timely intervention and improved treatment outcomes. Second, by analyzing the evolution of patient data in time, healthcare providers may tailor interventions, optimizing treatment efficacy and patient satisfaction. From an economic perspective, the improved efficiency of diagnosis and treatments may reduce overall healthcare costs. Finally, the developed models may be employed or fine-tuned in related data-scarce contexts.</p><p>As part of our investigation, we conducted a preliminary study on voice disorder detection, a key aspect of non-invasive diagnostics. We trained a deep learning model for analyzing voice recordings that achieves very high precision in detecting the presence of pathology and accurately identifies the type of pathology. We employed a transformer model <ref type="bibr" target="#b2">[3]</ref>, trained end-to-end (E2E) directly on the raw data, outperforming traditional methods such as convolutional neural networks (CNN) trained on frequency-transformed data. These preliminary results demonstrate the potential of our approach in enabling accurate and efficient non-invasive diagnostics for voice disorders.</p><p>The paper begins with an introduction to AI and its medical applications (Section 2). It then outlines the proposed idea, and it addresses foreseen challenges (Section 3). We then present the preliminary results on a voice disorder detection case in Section 4, and conclude with a discussion on the framework's impact (Section 5).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1:</head><p>Outline of the proposed pipeline. A user records different types of data, e.g., pictures, voice, texts, heart monitors, and sleep conditions. These data are uploaded to the cloud and processed by an artificial intelligence method. The results are visualized on a user-controlled application but can also be shared with a remote doctor who can require further exams.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>Deep Learning and Transformers Deep Learning is a robust method for uncovering patterns and insights from extensive datasets <ref type="bibr" target="#b3">[4]</ref>. Unlike conventional machine learning approaches, Deep Learning models learn directly from raw data in an E2E manner, without the need for manual feature engineering. Transformer models are a prominent class of Deep Learning architectures <ref type="bibr" target="#b2">[3]</ref>, demonstrating the effectiveness of this approach in analyzing sequential and multimodal data <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>. Unlike traditional recurrent and convolutional neural networks, Transformers can capture long-range dependencies and preserve contextual information over extended sequences, thanks to their self-attention mechanisms. This characteristic enables Transformers to process raw multimodal sequential data, such as text or time-series data, as well as images, making them highly adaptable for various applications, including medical diagnostics <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>DL for Medicine</head><p>Deep Learning has emerged as a transformative technology in various healthcare applications. First, DL can analyze electronic health records <ref type="bibr" target="#b8">[9]</ref>, enabling personalized treatment recommendations and predictive analytics for patient outcomes. In medical imaging <ref type="bibr" target="#b9">[10]</ref>, DL models can interpret radiological images, including X-rays, CT scans, and MRI images, with levels of accuracy comparable to or even surpassing that of human experts. These models have been utilized for tasks such as disease classification, lesion detection, and tumor segmentation, the latter also performed in realtime during surgery. In drug discovery, AI algorithms have been shown capable of developing novel drugs through accurate predictions of protein structures <ref type="bibr" target="#b10">[11]</ref>. Finally, DL models have also shown increased capability in handling multi-omics and multi-modal data <ref type="bibr" target="#b11">[12]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">AI for non-invasive medicine</head><p>In this paper, we propose a framework for analyzing an individual's health condition in time through noninvasive AI-powered diagnostics. Our proposed pipeline embodies a user-centric approach, empowering individuals to actively participate in their healthcare journey. From a medical point of view, we consider the analysis of all pathologies detectable by human experts through noninvasive diagnosis. Drawing parallels with advancements in other domains, we argue that AI models can replicate and potentially surpass human diagnostic accuracy also in this domain. This solution may enable large-scale screening through simple but accurate analysis of the patient data collected remotely in a non-invasive way.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Data collection</head><p>In Figure <ref type="figure">1</ref> we report a visualization of the proposed framework. Leveraging everyday devices such as smartphones, laptops, or wearables, users can effortlessly record diverse types of data. Clearly, each type of device enables the collection of different data types. Laptops and smartphones allow collecting typing and click patterns, texts, voice recordings, and pictures (the last two particularly through smartphones). Wearables, on the other side, allow (and are already employed for) monitoring heart rate, blood pressure, athletic performance and sleep conditions, to cite a few.</p><p>Data storage analysis Upon collection, the data are uploaded to a cloud-based database and analyzed by means of a single advanced artificial intelligence model. The employment of the aforementioned transformer models accommodate diverse data modalities and sources. By integrating multi-modal data processing capabilities, our framework aims to capture a holistic view of an individual's health status, enabling comprehensive and personalized diagnostics. Moreover, the collection of continuous health data enable taking into consideration in-time evolution and predictive modelling. The model may also evolve and improve over time by means of the new data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Result visualization</head><p>The processed results are then presented to users through an intuitive and usercontrolled application interface. Through this interface, users gain valuable insights into their health metrics, facilitating informed and proactive decisions regarding their healthcare. Furthermore, a user can transmit the processed data and diagnostic results to their remote healthcare professionals, allowing them to make informed clinical decisions and make timely interventions. Additionally, healthcare providers can utilize the data for population-level health monitoring and epidemiological studies, facilitating the identification of emerging health trends and proactive public health interventions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Challenges</head><p>The effectiveness of the proposed framework relies also on our ability to address and resolve a number of both technical and non-technical issues.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Data quality</head><p>The collection of data through personal devices introduces the risk of data noise, which encompasses various factors such as sensor inaccuracies, environmental interference, and incorrect user input. Sensor inaccuracies may lead to erroneous measurements, while environmental interference, such as background noise or lighting conditions, can distort the recorded data. Additionally, users may input incorrect data, such as taking pictures of the wrong part of their body, which can further exacerbate data noise and impact the analysis model. These challenges can hinder the model's ability to accurately interpret and analyze the data, potentially leading to erroneous conclusions and suboptimal performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Research direction: data augmentation + input data checking</head><p>To address the challenges posed by sensor inaccuracies and environmental interferences, a robust data augmentation pipeline can be implemented to mitigate these sources of noise. By incorporating various data augmentation techniques such as noise injection, signal filtering, and data synthesis, the pipeline can generate diverse and representative training data that encapsulates the variability present in real-world scenarios <ref type="bibr" target="#b12">[13]</ref>. Specific preprocessing tailored to each data modality can help normalize and enhance the quality of the collected data. Furthermore, to mitigate the issue of incorrect user input, a dedicated model can be employed to validate the accuracy of the collected data. This model can check each type of collected data, such as images or sensor readings, to verify their correctness and flag any discrepancies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Privacy preservation</head><p>The second problem arises from privacy concerns associated with collecting sensitive data and transmitting it to a cloud-based database. Individuals may be hesitant to share sensitive information due to concerns about data security and privacy breaches <ref type="bibr" target="#b13">[14]</ref>. Transmitting such data to a cloud-based database further exacerbates these concerns, as it involves relinquishing control over personal information to third-party service providers. Moreover, regulatory compliance <ref type="bibr" target="#b14">[15]</ref> impose stringent requirements on the handling of personal health information, adding complexity to the data collection process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Research direction: a federated learning approach</head><p>The privacy issue can be effectively addressed through the adoption of a federated approach. Federated learning is an area of machine learning studying how to coordinate and distribute training over several models without exchanging raw data. The goal is to mitigate privacy concerns associated with transmitting sensitive data to a centralized cloud-based database. Furthermore, personalized models generated through federated learning may exhibit higher diagnostic accuracy, as they are trained on data specific to each user's health profile.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.3.">Human understanding &amp; trustworthiness</head><p>The third issue concerns the explainability challenge associated with employing deep learning-based black box models for predictions. While these models offer impressive performance, their inherent complexity makes them opaque and difficult to interpret. This lack of explainability presents challenges for clinicians and end-users who require insights into the model's decision-making process <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b18">19]</ref>. Without transparent explanations, stakeholders and regulatory institutions may hesitate to trust the diagnostic and monitoring framework.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Research direction: Concept-based XAI models</head><p>A possible solution is represented by the employment of eXplainable AI (XAI) algorithms that can shed light on the decisions of the models <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22]</ref>. Particularly, Concept-based XAI models offer intrinsic interpretability by mapping raw data to interpretable high-level concepts before making class predictions, offering insights into the model's decision-making process <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25]</ref>. This approach not only addresses the explainability challenge but also fosters trust and confidence in the diagnostic and monitoring outcomes produced by the system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">The case of voice disorder detection</head><p>Vocal disorders are prevalent pathologies affecting a significant portion of the population and exerting a substantial impact on patients' quality of life <ref type="bibr" target="#b25">[26,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b27">28,</ref><ref type="bibr" target="#b28">29]</ref>. These disorders may originate from various causes, including both benign and malignant conditions, and neurodegenerative disorders <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b30">31,</ref><ref type="bibr" target="#b31">32]</ref>. Diagnosis often relies on clinicians' auditory assessments of patients' voices, highlighting the critical need for accurate and timely detection. Here, a DL model is used to analyze the raw recordings and automatically detect patterns indicative of vocal disorders and distinguish between various pathologies, including nodules, polyps, cysts, spasmodic dysphonia or vocal cord paralysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Preliminary experiments</head><p>We demonstrate significant advancements over prior attempts in voice disorder detection using AI models <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34]</ref>    accuracy, up to +30%, primarily attributed to the utilization of a Transformer model rather than a Convolutional Neural Network (CNN). The Transformer's inherent ability to process raw time-series data E2E without any timefrequency preprocessing offers distinct advantages in analyzing voice data <ref type="bibr" target="#b34">[35]</ref>. Additionally, we introduce a robust data augmentation pipeline and consider both vowel and sentence-based recordings, further enhancing performance by up to +10% compared to the employment of a standard Transformer model only. As reported in Figure <ref type="figure" target="#fig_1">3</ref>, similar considerations hold also for the classification of the macro-category pathology. These enhancements underscore the efficacy of our approach in the diagnosing vocal disorders.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussion</head><p>Overall, the proposed pipeline holds promise for improving healthcare delivery, leveraging AI to enable noninvasive diagnostics and monitoring. The medical team involved in the development of this pipeline also believes in its transformative impact from a social, technological, and economic point of view.</p><p>Social Impact By enabling individuals to record various types of data remotely using everyday devices, the proposed pipeline facilitates non-invasive diagnostic procedures, eliminating the need for expensive and invasive tests. The accessibility of the proposed pipeline extends beyond traditional healthcare settings, allowing individuals in remote or underserved areas to access diagnostic services conveniently. Individuals can be monitored remotely, allowing healthcare providers to track their health status in real-time and intervene promptly if abnormalities are detected <ref type="bibr" target="#b35">[36]</ref>. Early detection and intervention facilitated by the pipeline lead to improved prognoses for patients, as healthcare providers can initiate treatment at earlier stages of disease progression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Technological Impact</head><p>The proposed framework has significant technological implications. By leveraging the adaptable nature of the framework, models developed for one medical application can be readily deployed and tested in related contexts, accelerating the pace of medical research and innovation. As an example, the presented voice disorder detection model could be tested for neurodegenerative patients. Additionally, the framework enables the fine-tuning of models for specific cases, including those with limited data availability, such as rare diseases. Despite the scarcity of data in such contexts, the model can still generalize due to its original training on a larger and diverse dataset. Economic Impact From an economic point of view, the proposed pipeline has low operating costs due to the utilization of personal devices for data collection and cloud-based analysis. Additionally, the framework facilitates the collection of low-cost, virtuous-cycle data, enabling continuous monitoring and feedback loops that enhance the accuracy and effectiveness of diagnostic and monitoring processes over time. Moreover, the framework lowers the burden of diagnosis and monitoring on public healthcare structures. This, in turn, enables healthcare professionals to focus on more complex and critical medical issues, ultimately improving the efficiency and resource allocation of the healthcare system.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Comparative analysis of the test model performance in distinguishing Healthy individuals from those afflicted with a pathological condition.</figDesc><graphic coords="4,100.47,84.19,178.97,121.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Comparative analysis of test model performance in identifying the macro pathology that afflicts individuals.</figDesc><graphic coords="4,100.47,256.69,178.97,133.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>1 . As reported in Figure2, our approach achieves notable improvements in</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">We re-implemented these models for a fair comparison.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Ai in health and medicine</title>
		<author>
			<persName><forename type="first">P</forename><surname>Rajpurkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Banerjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Topol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature medicine</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="31" to="38" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Non-invasive interventions for improving well-being and quality of life in patients with lung cancer</title>
		<author>
			<persName><forename type="first">J.-R</forename><surname>Rueda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pascual</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Casacuberta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cochrane Database of Systematic Reviews</title>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ł</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Deep learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">nature</title>
		<imprint>
			<biblScope unit="volume">521</biblScope>
			<biblScope unit="page" from="436" to="444" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">wav2vec 2.0: A framework for self-supervised learning of speech representations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Baevski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Auli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="12449" to="12460" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Self-supervised speech representation learning by masked prediction of hidden units</title>
		<author>
			<persName><forename type="first">W.-N</forename><surname>Hsu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bolte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-H</forename><forename type="middle">H</forename><surname>Tsai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lakhotia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Salakhutdinov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hubert</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE/ACM Transactions on Audio, Speech, and Language Processing</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="3451" to="3460" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Transformers in medical image segmentation: A review</title>
		<author>
			<persName><forename type="first">H</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biomedical Signal Processing and Control</title>
		<imprint>
			<biblScope unit="volume">84</biblScope>
			<biblScope unit="page">104791</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">How much attention should we pay to mosquitoes?</title>
		<author>
			<persName><forename type="first">M</forename><surname>La Quatra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vaiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cagliero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Garza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<idno type="DOI">10.1145/3503161.3551594</idno>
		<idno>doi:10.1145/3503161.3551594</idno>
		<ptr target="https://doi.org/10.1145/3503161.3551594" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 30th ACM International Conference on Multimedia, MM &apos;22</title>
				<meeting>the 30th ACM International Conference on Multimedia, MM &apos;22<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="7135" to="7139" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Scalable and accurate deep learning with electronic health records</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rajkomar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Oren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hajaj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Marcus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NPJ digital medicine</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">18</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of deep learning in medical imaging</title>
		<author>
			<persName><forename type="first">K</forename><surname>Suzuki</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Radiological physics and technology</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="257" to="273" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Highly accurate protein structure prediction with alphafold</title>
		<author>
			<persName><forename type="first">J</forename><surname>Jumper</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Evans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pritzel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Green</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Figurnov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ronneberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tunyasuvunakool</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bates</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Žídek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Potapenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">596</biblScope>
			<biblScope unit="page" from="583" to="589" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A survey on data integration for multi-omics sample clustering</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lovino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Randazzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ciravegna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barbiero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ficarra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cirrincione</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">488</biblScope>
			<biblScope unit="page" from="494" to="508" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Transformer-based non-verbal emotion recognition: Exploring model portability across speakers&apos; genders</title>
		<author>
			<persName><forename type="first">L</forename><surname>Vaiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>La Quatra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cagliero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Garza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<idno type="DOI">10.1145/3551876.3554801</idno>
		<idno>doi:10.1145/3551876.3554801</idno>
		<ptr target="https://doi.org/10.1145/3551876.3554801" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd International on Multimodal Sentiment Analysis Workshop and Challenge, MuSe&apos; 22</title>
				<meeting>the 3rd International on Multimodal Sentiment Analysis Workshop and Challenge, MuSe&apos; 22<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="89" to="94" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Privacy and security issues in deep learning: A survey</title>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Vasilakos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="4566" to="4593" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Madiega</surname></persName>
		</author>
		<title level="m">Artificial intelligence act, European Parliament</title>
				<imprint>
			<publisher>European Parliamentary Research Service</publisher>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The importance of interpretability and visualization in machine learning for applications in medicine and health care</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vellido</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computing and applications</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Exploring subgroup performance in end-to-end speech models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pastor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Attanasio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Mazzia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Giollo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gueudre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cagliero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>De Alfaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amberti</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICASSP49357.2023.10095284</idno>
	</analytic>
	<monogr>
		<title level="m">ICASSP 2023 -2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="5" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Towards comprehensive subgroup performance analysis in speech models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pastor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Attanasio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Mazzia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Giollo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gueudre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Reale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cagliero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cumani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>De Alfaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amberti</surname></persName>
		</author>
		<idno type="DOI">10.1109/TASLP.2024.3363447</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE/ACM Transactions on Audio, Speech, and Language Processing</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Bad exoplanet! explaining degraded performance when reconstructing exoplanets atmospheric parameters</title>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giobergia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=9Z4XZOhwiz" />
	</analytic>
	<monogr>
		<title level="m">NeurIPS 2023 AI for Science Workshop</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Vilone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Longo</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2006.00093</idno>
		<title level="m">Explainable artificial intelligence: a systematic review</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Explaining speech classification models via word-level audio segments and paralinguistic features</title>
		<author>
			<persName><forename type="first">E</forename><surname>Pastor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Koudounas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Attanasio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hovy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<meeting>the 18th Conference of the European Chapter of the Association for Computational Linguistics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Logic explained networks</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ciravegna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barbiero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giannini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lió</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Maggini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Melacci</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">314</biblScope>
			<biblScope unit="page">103822</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Interpretable neural-symbolic concept reasoning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Barbiero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ciravegna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giannini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Espinosa</forename><surname>Zarlenga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Magister</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tonda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Precioso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jamnik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Marra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th International Conference on Machine Learning</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Krause</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Brunskill</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Cho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Engelhardt</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Sabato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Scarlett</surname></persName>
		</editor>
		<meeting>the 40th International Conference on Machine Learning<address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">202</biblScope>
			<biblScope unit="page" from="1801" to="1825" />
		</imprint>
	</monogr>
	<note>Proceedings of Machine Learning Research</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Concept embedding models: Beyond the accuracy-explainability trade-off</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Espinosa</forename><surname>Zarlenga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barbiero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ciravegna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Marra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giannini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Diligenti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Shams</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Precioso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Melacci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Weller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="21400" to="21413" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Poeta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ciravegna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pastor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Cerquitelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.12936</idno>
		<title level="m">Concept-based explainable artificial intelligence: A survey</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Voice disorders in the general population: prevalence, risk factors, and occupational impact</title>
		<author>
			<persName><forename type="first">N</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Merrill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">D</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Smith</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Laryngoscope</title>
		<imprint>
			<biblScope unit="volume">115</biblScope>
			<biblScope unit="page" from="1988" to="1995" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Self-reported impact of dysphonia in a primary care population: An epidemiological study</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Cohen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Laryngoscope</title>
		<imprint>
			<biblScope unit="volume">120</biblScope>
			<biblScope unit="page" from="2022" to="2032" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">The prevalence of voice problems among adults in the united states</title>
		<author>
			<persName><forename type="first">N</forename><surname>Bhattacharyya</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Laryngoscope</title>
		<imprint>
			<biblScope unit="volume">124</biblScope>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Voice disorders in the general greek population and in patients with laryngopharyngeal reflux. prevalence and risk factors</title>
		<author>
			<persName><forename type="first">N</forename><surname>Spantideas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Drosou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Karatsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Assimakopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Voice</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="389" to="e416" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Prevalence of benign vocal fold lesions: Long-term results from a single european institution</title>
		<author>
			<persName><forename type="first">E</forename><surname>Brunner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Eberhard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gugatschka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Voice</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Gradient boosting for parkinson&apos;s disease diagnosis from voice recordings</title>
		<author>
			<persName><forename type="first">I</forename><surname>Karabayir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Goldman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pappu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Akbilgic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">BMC Medical Informatics and Decision Making</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Voice-based classification of amyotrophic lateral sclerosis: where are we and where are we going? a systematic review</title>
		<author>
			<persName><forename type="first">H</forename><surname>Vieira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Costa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sousa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Reis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Coelho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurodegenerative Diseases</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="163" to="170" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Voice pathology detection using convolutional neural networks with electroglottographic (egg) and speech signals</title>
		<author>
			<persName><forename type="first">R</forename><surname>Islam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Abdel-Raheem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tarique</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Methods and Programs in Biomedicine Update</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">100074</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">A voice disease detection method based on mfccs and shallow cnn</title>
		<author>
			<persName><forename type="first">X</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ding</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Voice</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">End-to-End Neural Transformer Based Spoken Language Understanding</title>
		<author>
			<persName><forename type="first">M</forename><surname>Radfar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mouchtaris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kunzmann</surname></persName>
		</author>
		<idno type="DOI">10.21437/Interspeech.2020-1963</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. Interspeech 2020</title>
				<meeting>Interspeech 2020</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="866" to="870" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Real-time analysis of physiological data to support medical applications</title>
		<author>
			<persName><forename type="first">D</forename><surname>Apiletti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Baralis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bruno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Cerquitelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE transactions on information technology in biomedicine</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page" from="313" to="321" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
