=Paper=
{{Paper
|id=Vol-1683/hda16_li_x
|storemode=property
|title=Discovery of learning topics in an online social network for health professionals
|pdfUrl=https://ceur-ws.org/Vol-1683/hda16_li_x.pdf
|volume=Vol-1683
|authors=Xin Li,Karin Verspoor,Kathleen Gray,Stephen Barnett
}}
==Discovery of learning topics in an online social network for health professionals==
Discovery of learning topics in an online social network for health professionals Xin LIa,1, Karin VERSPOORa, Kathleen GRAYa and Stephen BARNETTb a Health and Biomedical Informatics Research Centre, University of Melbourne b General Practice Academic Unit, Graduate School of Medicine, University of Wollongong Abstract. Online social networking provides support to health professionals' learning and professional development. To understand their learning needs in this context, this study employs topic modelling of postings to an online social network for health professionals to identify the topics of interest. The analysis shows that the health professionals in this network were more interested in discussing non- clinical topics than clinical ones. The non-clinical topics include some controversial topics such as policy-related issues, as well as an interest in the latest news and advanced information in the field. The clinical topics relate to their practices, including sharing practical and experiential knowledge and providing benchmarks. Keywords. Topic modelling, networked learning, health professional education 1. Introduction As medical knowledge expands and healthcare delivery becomes more complex, health professionals must commit to continuous learning to maintain up-to-date knowledge and skills. One approach to meeting their learning and development needs is through engagement in an online social network (OSN) [1]. OSNs have been found useful to reduce professional isolation and support anytime-anywhere peer-to-peer interaction at scale. Also, they are thought to contribute to the development of professional networks and improve continuing professional development. There are many OSN targeted towards health professionals but they appear to fail to support the broader learning objectives [2]. It has been recognised that there is a lack of understanding about how health professionals learn in an OSN, making it difficult to design and facilitate this type of learning [3]. To realise the full potential of OSNs for health professionals’ learning, understanding and evaluating this learning context is important. Previous studies focused on understanding learning behaviours by identifying the patterns of the interaction among health professionals [4, 5]. However, there is still much to be explored in terms of the textual dialogue among health professionals, particularly regarding how those dialogues support the process of learning. This paper proposes topic modelling as a method to discover the topics of interest from an OSN for health professionals. The identified topics can provide insights on the learning resource and professional development needs of the health professionals. 2. Background and Related Work Previous work has been done on analysis of dialogue in online learning environments to find evidence about learning and knowledge construction. This has required considerable resources and effort for manual data coding to analyse cognitive and social processes in which learners engage. For example, De Laat [6] assessed the quality of the dialogue in an online community for the police using a coding scheme that examines the social construction of knowledge. Schrire [7] investigated the 1 Corresponding Author: Xin Li; E-mail: xinli87@gmail.com. knowledge-building process in a discussion forum used in a higher education context by applying community of inquiry model. As more and more textual data is generated online and human annotation becomes impossible, computational tools such as topic modelling become more useful. Topic modelling is a statistical method that analyses the words of the original texts to discover the themes that run through them, how those themes are connected to each other, and how they change over time [8]. Many researchers have used topic modelling to explore the themes in dialogues in online learning environments. However, to the best of our knowledge, its application in discovering topics among the online community of health professionals is novel in health professional education research. Tobarra, Robles-Gómez [9] used it to discover topics of interest in the forum of a Learning Management System for improving the structure and contents of education courses. Portier, Greer [10] used it together with sentiment analysis to identify improvements that enhance social support in an online cancer community. Most recently, Ezen-Can, Boyer [11] used it to understand the topics of discussion in the forum of an open online course for educators. 3. Methods 3.1. Dataset Data were collected from the database of an online discussion forum provided by a health professional OSN host organisation, with Human Research Ethics approval. The online forum was established in 2009 specifically for registered health practitioners and had more than 10,000 members. Since the online forum was set up for doctors to discuss industry issues, share best practices and promote conversation within the health community, it is logical to assume that the topics discovered from the forum posts would reflect the resource and professional development needs of this community. The data for this study comprised all the posts made by the forum participants (N = 48) who remained active in three consecutive years from the period 2012 to 2014. The three-year period represents 50% of the overall operating period of this forum, and the most recent and complete years available at the time of data collection in 2015. The 48 forum participants represent 13% of overall participants during this period. 154 discussion threads were found, each receiving between one and 58 replies. A total of 1604 posts (105,063 words) were extracted from the forum. 3.2. Topic Modeling Using MALLET To identify the topics of interest in this forum, we generated a topic model using the MALLET tool implemented in R. The MALLET (Machine Learning for Language Toolkit) automates the process of topic discovery from a large volume of text; it implements the latent Dirichlet allocation (LDA) algorithm, which is a generative probabilistic model [12]. The basic idea of LDA is that documents are presented as a random mixture of topics, where each topic is a probability distribution over a given vocabulary of words [13]. In this study, a document is defined as a forum post. The MALLET program was used to generate clusters of words (i.e. topics) that frequently occur together within a forum post. 3.3. Procedure Data preparation: We pulled full text from each post using SQL queries, and cleaned the text by removing anything other than English letters or spaces. To improve the coherence of generated topics, we removed the stop words from the full text based on the standard list of stop words of MALLET2. We also further removed popular words (e.g. lol, cheers, pretty, nice, yrs) and any specific words associated with country/state/city and personal names (e.g. sherlock, watson, judas) that appeared in 2 For instructions to download the standard list of stop words of MALLET, please go to http://mallet.cs.umass.edu/import-stoplist.php. this dataset. In addition, all words were stemmed to retrieve their stems so that various forms of a word would be counted together when counting word frequency. This was done using the stemmer function (available in tm package) in R. These pre-processing steps reduced the number of words in the dataset to 54873. Topic model generation: To generate topic models using MALLET, two variables (i.e. number of topics, number of sampling iterations) must be defined. To identify the optimal number of topics for the topic model, we specified different numbers of topics to generate four models (Models 1 – 4). The initial number of topics was set to 15 by inspecting all the 154 thread titles and noting from inspection that there are approximately 14 broad topics in the dataset. The dataset of this size usually has the default sampling iteration set to 400. Since increasing the number of iterations may improve topic coherence [12], we increased the iteration to 800 to generate two additional models (Models 5 – 6). Table 1 depicts the variables defined for these different topic models. Table 1. The variables of various topic models Model 1 Model 2 Model 3 Model 4 Model 5 Model 6 Number of Topics T 15 20 25 30 20 25 Number of Iterations I 400 400 400 400 800 800 Topic inference: Topics were inferred using clusters of words produced by topic models. Since each topic is a probability distribution over words, we chose to inspect the top ten words for each cluster. This is based on the assumption that more words per cluster might make it more difficult to infer a meaningful topic for each cluster. Topic optimisation: Inferred topics were optimised by reviewing the contents of the top five posts with consideration for each word cluster (i.e. topic). The top five posts for each topic were identified by inspecting the probability of each topic appearing in each post, which was obtained by employing the function mallet.doc.topics in MALLET. The optimisation helped identify further duplicates and improve the accuracy of the inferred topics. Topic evaluation: Traditionally, the performance of topic models are typically evaluated using quantitative intrinsic methods such as computing the probability of held-out documents. However, it has been shown that this measure is not always a good predictor of human judgment [14]. In this study, we evaluated the topics based on human judgment using F-measure [15], which is often used in the field of information retrieval. There are four performance metrics considered (i.e. accuracy, precision, recall, and F-score, as defined in Table 2). Table 2. The performance metrics Performance Description Formula metric Accuracy The percentage of the posts identified are TP + TN / (TP + FP + FN + TN) expected to belong to an optimised topic. Precision The percentage of posts correctly identified as TP / (TP + FP) belonging to an optimised topic. Recall The percentage of posts identified as belonging TP / (TP + FN) to any topic. F-score The harmonic mean of precision and recall, 2 × (Precision × Recall) / (Precision which can be interpreted as a weighted average + Recall) of the Precision and Recall. We randomly selected 40 forum posts from the dataset to validate the optimised topics using F-measure. We considered that a post is True Positive (TP) when any part of the post content matches an optimised topic; a post is False Positive (FP) when the post content does not match an optimised topic; a post is False Negative (FN) when the post content suggests a discernable topic (may be an optimised topic or any new topic that has not been identified); a post is True Negative (TN) when the post content does not suggest any discernable topic. 4. Results and Discussion 4.1. Topic Model Comparison After inferring the topics of generated topic models, we compared the number of optimised topics from each topic model. It seems that a topic model with T = 20 would be more appropriate than T = 15, or T = 25, or T = 30. As shown in Table 3, Model 1 (T = 15) generated 9 optimised topics which indicate that setting too few topics could result in not covering all topics. Model 3 (T = 25) generated 11 optimised topics which indicate that setting too many topics could result in duplications (five pairs of word clusters represent the same topic). Model 4 (T = 30) generated only 9 optimised topics which indicate that setting too many topics could even result in uninterpretable topics. For this dataset, a topic model with I = 400 would be more appropriate than I = 800. The number of optimised topics generated from Model 5 and Model 6 suggests that increasing the number of iterations did not result in better topic models, as the composition and quality of the resulting topics only increased to a certain point and then levelled off. From the results, we concluded that Model 2 (T = 20, I = 400) seems to be the topic model that best describes the topics of interest discussed by health professionals in this forum. Table 3. Number of inferred and optimised topics for various topic models Model 1 Model 2 Model 3 Model 4 Model 5 Model 6 Number of inferred topics 11 14 15 13 13 12 Number of optimised topics 9 13 11 9 12 9 4.2. Optimised Topics in the Selected Topic Model Table 4 shows the inferred, optimised topics and their associated word clusters for the selected topic model (i.e. Model 2). Of the 20 word clusters generated from the model, 14 unique topics were inferred; two word clusters were interpreted as referring to the same topic, and four word clusters were indicated as “Not Applicable” (N.A.) as no meaningful topic could be inferred. After optimising the inferred topics by reviewing the selected posts, 13 unique topics were found. They are identified as a mixture of clinical and non-clinical topics. The clinical topics (CT) include “palliative care”, “rheumatology”, “evidence-based medicine”, “statins use”, “vitamin use”, “vaccines”, “women’s health check”, and “fibromyalgia”; the non-clinical topics (NCT) include “patient fees”, “training”, “prescriptions”, “policy”, and “workload”. Table 4. The optimised topics for the selected model Topic Cluster Topic words Inferred topic Optimised topic weight pay medicare bulk work charge fee 1 0.19594 Bulk billing Patient fees (NCT) private service money government prescription script pharmacy addict drug Prescriptions 2 0.18415 Prescriptions pharmacist pbs authority day write (NCT) food car house fridge eat poor store hot 3 0.0322 N.A. N.A. change balance patient pay medicine stress society 4 0.1898 Health cost Patient fees (NCT) reduce current government rate finance care nurse hour visit service palliative 5 0.08499 Palliative care Palliative care (CT) home medical provide rural medical profession racgp public doctor 6 0.04848 Training Training (NCT) ahpra nurse health wrote report point vaccine medicine understand body 7 0.05038 Vaccines Vaccines (CT) view generate form base suggestion refer comment expert issue interest 8 0.03467 N.A. N.A. person lack call present programme restrict country year work moratorium 9 0.07868 Policy Policy (NCT) hospital system area dws law trial statin evidence effect prevent group 10 0.10449 Statins use Statins use (CT) side benefit interest study effect level side dose disease vitamin 11 0.07429 Vitamin use Vitamin use (CT) high symptom difference drug pain inject joint muscle guidance bursa Elbow bursa Rheumatology 12 0.09575 knee stretch tear elbow treatment (CT) evidence point medicine base comment Evidence- Evidence-based 13 0.07989 treatment understand body view ahpra based medicine medicine (CT) examination check breast pap women Women’s Women’s health 14 0.07777 history year present cancer diagnose health checks checks (CT) human organ end therapy cell central 15 0.1173 N.A. N.A. protect age create state risk diabetes cholesterol disease level 16 0.0878 Statins use Statins use (CT) statin calculate exercise hdl year pain chronic toe relief chest attached 17 0.03857 Fibromyalgia Fibromyalgia (CT) post bit metal fibromyalgia patient practice doctor person time 18 0.17284 Work hours Workload (NCT) medical work care health hour practice hospital general training 19 0.06752 Training Training (NCT) specialist nurse year medical head base record request summary advice provide 20 0.02064 N.A. N.A. inform letter email initial legal As shown in Table 4, there are more clinical than non-clinical topics identified from the dataset. However, the weights of the topics imply that non-clinical topics were more frequently discussed than clinical ones. With regards to the clinical topics, palliative care, rheumatology, and evidence- based medicine appeared to generate some in-depth discussion among the participants. By inspecting a number of specific posts on the topics relating to women’s health checks, fibromyalgia, the use of statins, vaccines, and vitamin, we noted that the participants were interested in benchmarking their practices. This is understandable as clinical practice can be conducted differently in different places; OSNs have been found to enable health professionals to share different ways of performing the same practice and benchmark the most effective one [16]. The non-clinical topics identified from this dataset are mostly controversial (include policy, workload and patient fees). This finding is consistent with previous studies that have demonstrated health professionals are particularly interested in discussing controversial topics in an OSN [17]. In addition, the participants were keen to keep themselves up-to-date on advanced information and news in the field; this is reflected in the topics relating to policies, training, and prescriptions. 4.3. Topic Evaluation The 13 identified topics were evaluated using F-measure against 40 randomly selected posts from the dataset. The Accuracy, Precision, Recall, and F-score of the topic model were 0.53, 0.63, 0.70, and 0.66 respectively. The Accuracy of 0.53 indicates that the topic model is likely to capture 53% of the topics in any randomly selected posts. The F-score of 0.66 informs that the topic model correctly captures 66% of the overall topics in this random selection of 40 posts. 5. Conclusions and Future Work OSNs have been increasingly used by health professionals to share medical knowledge and experience. However, there is a lack of understanding about how health professionals learn in OSNs, making it difficult to design and facilitate this type of learning. This study contributes towards understanding their learning resource and development needs in OSNs by demonstrating the use of topic modelling to identify the topics of interest that emerge from an online discussion forum of health professionals. The evaluation of the topic model was performed using F-measure. The F-score of 0.66 informs that the topic model is not optimal but correctly captures 66% of the overall topics in a random selection of 40 posts. This suggests that topic modelling could be used to identify the emerging learning topics from the large amount of textual dialogue generated in an OSN. As we have found no previous work on topics discussed by an OSN for health professionals to compare our results with, it is inconclusive whether the topics we identified are typical or atypical of those discussed by health professionals. However, the results suggest that the health professionals in this OSN are interested in knowing or discussing clinical topics relating to palliative care, rheumatology, evidence-based medicine, women’s health checks, fibromyalgia, the use of statins, vaccines, and vitamins, as well as non-clinical topics relating to prescriptions, patient fees, policy, workload, and training. Identifying topics using this method could provide education designers and OSN operators with guidance on facilitating online discussion that is most relevant to the learning needs of health professionals. In this OSN, it has been found that non-clinical topics were more frequently discussed than clinical ones by the health professionals. Without knowing the context, we could not support having non-clinical topics as the main focus of their online discussion, but it is important to consider how to help health professionals to deal with the challenge of keeping themselves up-to-date on non- clinical and work-related information. In addition, it might be worth considering proposing common clinical topics relating to their clinical practices that allow them to share practical and experiential knowledge and meet the needs for benchmarking. A limitation of this study is that considering the overall activity in the discussion forum within this OSN, data were analysed very selectively. Due to limitations of the data source, passive users (i.e. those who learn by reading but do not participate in any discussion) were not tracked in our study, which means the topics identified only apply to the active participants of this OSN. In a future study, we plan to include additional meta-data to fit into the topic model, for example, including the identity of the authors enables us to investigate author similarity based on their discussion of topics. This will help to group health professionals who may have similar learning needs. Furthermore, understanding of the learning context (e.g. goals, tasks, preference, interests, and constraints) enhances the interpretation of the identified topics. 6. References 1. Cheston, C.C., T.E. Flickinger, and M.S. Chisolm, Social media use in medical education: a systematic review. Academic Medicine, 2013. 88(6): p. 893-901. 2. Sandars, J., P. Kokotailo, and G. Singh, The importance of social and collaborative learning for online continuing medical education (OCME): directions for future development and research. Med Teach, 2012. 34(8): p. 649-652. 3. Institute of Medicine, Redesigning Continuing Education in the Health Professions. 2010, National Academies Press: Washington, DC. 4. Stewart, S.A. and S.S.R. Abidi, Using Social Network Analysis to Study the Knowledge Sharing Patterns of Health Professionals Using Web 2.0 Tools. Biomedical Engineering Systems and Technologies, 2013. 273: p. 335-352. 5. Li, X., et al., Analysing Health Professionals' Learning Interactions in Online Social Networks: A Social Network Analysis Approach, in Health Informatics New Zealand Conference. 2015: Christchurch, New Zealand. 6. De Laat, M., Network and content analysis in an online community discourse, in Proceedings of the Conference on Computer Support for Collaborative Learning: Foundations for a CSCL Community. 2002, International Society of the Learning Sciences: Boulder, Colorado. p. 625-626. 7. Schrire, S., Knowledge building in asynchronous discussion groups: Going beyond quantitative analysis. Computers & Education, 2006. 46(1): p. 49-70. 8. Blei, D.M., Probabilistic topic models. Communications of the ACM, 2012. 55(4): p. 77-84. 9. Tobarra, L., et al. Discovery of interest topics in web-based educational communities. in Computers in Education (SIIE). 2012. Andorra la Vella: IEEE. 10. Portier, K., et al., Understanding topics and sentiment in an online cancer survivor community. JNCI Monographs, 2013. 47: p. 195-198. 11. Ezen-Can, A., et al. Unsupervised modeling for understanding MOOC discussion forums: a learning analytics approach. in Proceedings of the Fifth International Conference on Learning Analytics And Knowledge. 2015. ACM. 12. McCallum, A.K., MALLET: A Machine Learning for Language Toolkit. 2002. 13. Blei, D.M., A.Y. Ng, and M.I. Jordan, Latent Dirichlet Allocation. The Journal of Machine Learning Research, 2003. 3: p. 993-1022. 14. Chang, J., et al. Reading tea leaves: How humans interpret topic models. in Advances in neural information processing systems. 2009. 15. Van Rijsbergen, C.J., Information retrieval. 2nd ed. 1979, Butterworth. 16. Millar, B., K. Ho, and A.-M. Carvalho, Social media to support physician practice and CPD: Opportunities, issues, and an emergency medicine case study. BCMJ, 2016. 58(2): p. 94-96. 17. Panahi, S., J. Watson, and H. Partridge, Social media and physicians: exploring the benefits and challenges. Health Informatics Journal, 2014: p. 1460458214540907.