=Paper=
{{Paper
|id=Vol-2068/wii3
|storemode=property
|title=Supporting Real Estate Search Through Automatic Information Suggestion
|pdfUrl=https://ceur-ws.org/Vol-2068/wii3.pdf
|volume=Vol-2068
|authors=Goro Otsubo
|dblpUrl=https://dblp.org/rec/conf/iui/Otsubo18a
}}
==Supporting Real Estate Search Through Automatic Information Suggestion==
Supporting Real Estate Search Through Automatic Information Suggestion Goro Otsubo LIFULL Co., Ltd. Tokyo, Japan ohtsubogoro@lifull.com ABSTRACT in most cases the user does not know what type of detailed Searching real estate property is uncommon task for most of search conditions are available and how to set them. As a the users. As a result, the user is not familiar with the detailed result, they are not used effectively in the real estate search search condition which is useful for search. In this paper, we process. propose to use voice recognition as a support for real estate To solve this issue, we assumed that we need support system property search. First user set vague search condition with which can recommend proper detailed search condition to GUI. Then system listens conversation between users. From assist real estate search. To do that, we have chosen to utilize the conversation, system extracts keyword and suggest detailed the conversation between users. Asking detailed demand for search condition and real estate information search results with new real estate property is sometimes intimidating for the user. those conditions. We will discuss system design, algorithm Moreover, the user may not necessary know the proper word to used to link spoken words and detailed search condition and search the real state information. Therefore, we assumed that preliminary test results. the system which can extract detailed search condition from ACM Classification Keywords the casual user conversation would be effective to support the H.5.m. Information Interfaces and Presentation (e.g. HCI): real estate property search. Miscellaneous; See http://acm.org/about/class/1998/ for the Interfaces for searching information using voice have been full list of ACM classifiers. This section is required. studied for a long time, and commercial service which uti- lizes voice interface become popular in recent years due to Author Keywords popularization of smartphones and home devices. Recently Voice Recognition; Multi modal; GUI speech recognition on the server side has become widespread, the accuracy of speech recognition has dramatically improved INTRODUCTION even for unspecified speakers. In Japan, people will move to new real estate property only six times for the lifetime on average. In other words, searching However, in most cases speech recognition is used for simple real estate property is "An event that happens once in a decade". web search and search in the app store[10]. This is due to the As a result it is very difficult to become familiar with special fact that even though the precision of speech to text conversion terms which are used to search real estate property. has improved, the next step which is "understanding meaning" remains as tough problem yet. According to the research In recent years real estate information sites on the Internet have by Luger et al.,[8] although the user’s expectation for speech been widely used. Using those sites, the user can access and recognition is high, current speech recognition interface makes search real estate information by themselves. However, search the user feel very stressful when using it. interface of those sites are like query interface of the property database. They have every function available to search the Considering technical difficulty described above, we have cho- database, but they are not not necessarily easy to use for all sen to use voice recognition as a support role. The user is the user who has interest in real estate property. not assumed to directly talk to the search interface. Instead, system listens the conversation between users. The system Typical user can set the vague request to search the real estate recognizes the keyword in the conversation, and tries to ex- property like the town they want to search, layout of property plore the appropriate detailed search condition to search the and price. Further, for example, it is also possible to set real estate property. Matching spoken word by the user and detailed search conditions such as "Pets allowed". However, detailed search condition is the first challenge in this research. The frustration that the user feels when using the voice inter- face is that the system can not understand the meaning of the word they have pronounced. We need to match appropriate detailed search condition to words in users’ conversation. There are various attempt to incorporate speech interface in search as a support role for exploratory search. Andolina ©2018. Copyright for the individual papers remains with the authors. et al. proposed systems [4][5] that extract keywords from Copying permitted for private and academic purposes. WII’18, March 11, 2018, Tokyo, Japan usersâĂŹ conversations to stimulating human creative think- solve the word mismatch problem, many different approaches ing.We assumed similar approach could be effective in real have been proposed. estate search. In this research, we tried to utilize Word2Vec [9]. By using Other reason why an existing system using speech recognition Word2Vec, each word is vectorized and similarity between makes a user feel frustrated is that the system is not transparent words can be calculated. Basic idea in this research is that we [8]. In other words, the user has no way of knowing why the tried to match spoken words and detailed search condition not system returned a response when the misunderstood answer with simple word match, but with match between related words came back. For this reason, we need to make matching pro- expanded from spoken words and detailed search condition cess between spoken and recognized word and recommended using Word2Vec. Following is the algorithm we used. detailed condition as clear as possible. By showing every keywords searched during the detailed search condition and 1. For each detailed search condition, manually set two to five relation between them, we will be able to achieve that goal. related keywords. Also, attempt to avoid keyword input by selecting and manip- 2. For each manually set keyword, search related words us- ulating suggested keywords by touch has been proposed[7] . ing Word2Vec, and record them as "expanded related key- By using similar interaction, we assumed that we will be able words". to increase the effectiveness of the system. Even if system can not recognize the user’s intention correctly, user may be able 3. Convert conversation voice to text string using speech to to find and select interesting keywords displayed on the screen text conversion function. and explore related detailed search condition. We will describe the design of the system below. 4. Extract noun and verb from the recognized text, and record them as "spoken words" SYSTEM DESIGN 5. For each spoken words, search related words using Screen shot of developed system is shown in Figure 1. Word2Vec, and record them as "expanded related key- User can set the search conditions which are area to search, words". layout and price of real estate property via GUI. After set- ting the search condition, users have free conversation about 6. Tries to find same word from expanded keywords from their intention for new house between themselves. System spoken words, and expanded keywords from manually set continuously monitors their conversation and recognized text keywords. If same word can be found, put link between is shown in the lower part of the screen. System automati- spoken word and detailed search condition. cally and continuously extract keywords from conversation Example of matching is shown in Figure 2. and tries to find the related detailed search condition such as "Within 800 meters from convenience store", "Pet allowed". Next we will discuss the training data set for Word2Vec. Re- System also display the search result of the real estate property lated words extracted using Word2Vec depends on nature of specifying each detailed condition. Link is shown between training data set. As an example, we will show the result of re- detailed search condition and searched property. There are two trieved related words of the word "Noise" using Wikipedia[3] major challenges in developing the proposed system. First, as training data in Table 1.(Original words are in Japanese) we needed to develop algorithm which search detailed search Generally, words used in connection with "Noise" are lined condition using user’s conversation data. Second, we need in- up, but they are different from what the user associates when teraction interface which will effectively support user to search searching for a real estate property. Users who are concerned real estate information even if the system does not recognize about "Noise" may choose "Top floor" detailed search condi- the user’s intention correctly. We will discuss these challenges tion if they are concerned about the noise from the floor above. next. Users who care about the noise from the roads may choose "Higher than the second floor". In either case, those detailed MATCHING ALGORITHM BETWEEN DETAILED SEARCH search conditions has no relations to the words extracted using Wikipedia as a training data. CONDITION AND SPOKEN WORDS One of the major frustrations felt using existing speech inter- We have also gathered text data from the web site called All face is that the system recognizes only programmed keywords About Japan [1] which has large amount of text related to real while there is no clue for the user about which word to speak. estate property. We found out that related words extracted As a result, quite often system does not recognize the word is more suitable for user’s intuition when searching the real that the user pronounced. Even though the user pronounces the estate property. Retrieved related words of the word "Noise" word which has similar meaning to the keyword that system using All About Japan as training data is also shown in Table 1. recognizes, system can not understand the similarity between "Sound leak" is what user may be interested in when searching those words. real estate property. Similar difficulty is recognized in the field of question and However, number of words in All About Japan is not necessary answer retrieval(herein Q&A retrieval) task[6] . Major chal- enough. As a result, in many cases we could not extract related lenge for Q&A retrieval is word mismatch between the user’s words because spoken word does not exist in All About Japan question and the question-answer pairs in the archive [11]. To data set. Figure 1. Screen shot of developed system Rank Wikipedia All About Japan The other aspect that we should consider is transparency of 1 Soot Water leakage the system which is described earlier. To ease user frustration, 2 Air pollution Shaking we need to make inference process clear to the user. 3 Salinity Suspicious person Considering these factors, we designed the interaction inter- 4 Exhaust gas Sound leak face shown in Figure 1. Not only detailed search condition 5 Interference Exhaust gas and spoken words are displayed, but also manually set key- Table 1. Related words of "Noise" searched using different training data words and expanded keywords are shown on the screen. All the words displayed on the screen can be used as an search keyword by dragging and dropping the word into the text area shown in the lower part of the screen. Considering characteristics of each data set described above, As stated before, precision of voice-to-text speech and search we decided to use both data set for keyword extraction. First of detailed search condition is not necessary high. In that case, we tries to extract keywords from All About Japan data set. If user may be frustrated if we only display "no results found" no matching word is found, we tries to extract keywords using or totally irrelevant result. In this research we tried to display Wikipedia data set. as much words as possible on the screen. By viewing those words, it it probable that some words on the screen might be interesting for the user. If so, user can start new search via interaction on the display, not by voice recognition. USER INTERACTION By using the algorithm described above, we can expect that we will be able to search the detailed search condition better EVALUATION than simple word match. However, still there are errors and we We have conducted two types of evaluation so far. First, we don’t expect that we can search the detailed search condition evaluated how effective search algorithm is for various user with high precision. input. Since we would like to evaluate effectiveness of our Figure 2. Concept diagram to match spoken words and detailed search condition algorithm, we used text input rather than voice input to avoid to use the voice interface as an main interface. Even though error caused by speech recognition. current system only extracts and search with detailed search condition, the user tried to search with voice sentence like We have chosen fifty sentences from Q&A site about real "Search property in Tokyo area". Even though we designed estate search[2] such as "Good view and well-ventilated" . All the system so that the user can set those search condition with the sentences do not include the exact word in detailed search GUI, the user often forget that. condition. Therefore, none of the conditions can be searched using simple word match algorithm. For each sentence, we From the result of user test, we realized that interaction design have selected corresponding detailed search condition. System should be improved so that user does not misunderstand that displays up to four detailed search condition. If at least one system can recognize every request that user might have. In search condition displayed is related to input sentence, we the current design, search condition setting GUI is hidden evaluated it as a success. while the system listen the conversation. In this case, the user expects system can understand any word what they say. As a result, 34 out of 50(68%) sentences can be evaluated as To avoid such a misunderstanding, we need to show search success. 9 out of 34 sentences which are evaluated as success condition setting GUI upfront. When the user speaks, and do include manually defined extend keywords. If we remove not operate the GUI, we can show current voice recognition them from the result, 25 out of 41 (61%) sentences can be interface over the search condition setting GUIs. searched successfully using proposed algorithm. Second, using current system, we conducted simple user eval- uation. We have explained system’s concept and function, and put the system beside, we had conversation about what type CONCLUSION AND FUTURE DIRECTION of real estate property the user is interested in. Five users have We have developed the real estate search system which rec- participated in the test. During the user test, we got consistent ommends detailed search conditions from users’ conversation. response from the user. Every participants see the importance Evaluation result shows the mixed results. We could confirm of system’s assistant role. In some cases, the user can find the the potential of proposed algorithm. However, we also recog- interesting results. Four out of five participants see detailed nized that user interface design should be improved. Based on search condition that they have never searched real estate prop- the evaluation results, we will redesign the system interaction, erty with before. We also observed that the user quite often try and will conduct further user test to evaluate how effective system can support the search of real estate information. REFERENCES 7. Khalil Klouche, Tuukka Ruotsalo, Diogo Cabral, 1. 2017. All Abount Japan : House,Real estate property. Salvatore Andolina, Andrea Bellucci, and Giulio Jacucci. https://allabout.co.jp/r_house/. (2017). 2015. Designing for Exploratory Search on Touch Devices. In Proceedings of the 33rd Annual ACM 2. 2017. OKWeb-Sumai(Housing). Conference on Human Factors in Computing Systems https://okwave.jp/c622.html. (2017). (CHI ’15). ACM, New York, NY, USA, 4189–4198. DOI: 3. 2017. Wikipedia (Japanese). https://ja.wikipedia.org/. http://dx.doi.org/10.1145/2702123.2702489 (2017). 8. Ewa Luger and Abigail Sellen. 2016. "Like Having a 4. Salvatore Andolina, Khalil Klouche, Diogo Cabral, Really Bad PA": The Gulf Between User Expectation and Tuukka Ruotsalo, and Giulio Jacucci. 2015. Experience of Conversational Agents. In Proceedings of InspirationWall: Supporting Idea Generation Through the 2016 CHI Conference on Human Factors in Automatic Information Exploration. In Proceedings of Computing Systems (CHI ’16). ACM, New York, NY, the 2015 ACM SIGCHI Conference on Creativity and USA, 5286–5297. DOI: Cognition (C&C ’15). ACM, New York, NY, USA, http://dx.doi.org/10.1145/2858036.2858288 103–106. DOI: 9. Thomas Mikolov. 2013. word2vec:Tool for computing http://dx.doi.org/10.1145/2757226.2757252 continuous distributed representations of words. (2013). 5. Salvatore et al Andolina. 2015. IntentStreams: Smart https://code.google.com/word2vec/. Parallel Search Streams for Branching Exploratory 10. Verto. 2017. WhatâĂŹs the Future of Personal Assistant Search. In Proceedings of the 20th International Apps? (2017). http://research.vertoanalytics.com/ Conference on Intelligent User Interfaces (IUI ’15). what-the-future-of-personal-assistant-apps-webinar-deck. ACM, New York, NY, USA, 300–305. 11. Xiaobing Xue, Jiwoon Jeon, and W. Bruce Croft. 2008. 6. Jiwoon Jeon, W. Bruce Croft, and Joon Ho Lee. 2005. Retrieval Models for Question and Answer Archives. In Finding Similar Questions in Large Question and Answer Proceedings of the 31st Annual International ACM SIGIR Archives. In Proceedings of the 14th ACM International Conference on Research and Development in Information Conference on Information and Knowledge Management Retrieval (SIGIR ’08). ACM, New York, NY, USA, (CIKM ’05). ACM, New York, NY, USA, 84–90. DOI: 475–482. DOI: http://dx.doi.org/10.1145/1099554.1099572 http://dx.doi.org/10.1145/1390334.1390416