EconPapers    
Economics at your fingertips  
 

Multimodal Information Fusion for the Prediction of the Condition of Condominiums

Miroslav Despotovic, David Koch, Matthias Zeppelzauer, Stumpe Eric, Simon Thaler and Wolfgang A. Brunauer

ERES from European Real Estate Society (ERES)

Abstract: Today's data analysis techniques allow for the combination of multiple different data modalities, which should also allow for more accurate feature extraction. In our research, we leverage the capacity of machine learning tools to build a model with shared neural network layers and multiple inputs that is more flexible and allows for more robust extraction of real estate attributes. The most common form of data for a real estate assessment is data structured in tables, such as size or year of construction, but also descriptions of the real estate. Other data that can also be easily found in real estate listings are visual data such as exterior and interior photographs. In the presented approach, we fuse textual information and variable quantity of interior photographs per condominium for condition assessment and investigate how multiple modalities can be efficiently combined using deep learning. We train and test the performance of a pre-trained convolutional neural network fine-tuned with variable quantity of interior views of selected condominiums. In parallel, we train and test the pre-trained bidirectional encoder-transformer language model using text data from the same observations. Finally, we build an experimental neural network model using both modalities for the same task and compare the performance with the models trained with a single modality. Our initial assumption that coupling both networks would lead to worse performance compared to fine-tuned single-modal models was not confirmed, as we achieved the better performance with the proposed multi-modal model despite the impairment of a very unbalanced dataset. The novelty here is the multimodal modeling of variable quantity of real estate-related attributes in a unified model that integrates all available modalities and can thus use their complementary information. With the presented approach, we intend to extend the existing information extraction methods for automated valuation models, which in turn would contribute to a higher transparency of valuation procedures and thus to more reliable statements about the value of real estate.

Keywords: Avm; Computer vision; Hedonic Pricing; NLP (search for similar items in EconPapers)
JEL-codes: R3 (search for similar items in EconPapers)
Date: 2023-01-01
New Economics Papers: this item is included in nep-ain, nep-big, nep-cmp, nep-net and nep-ure
References: Add references at CitEc
Citations:

Downloads: (external link)
https://eres.architexturez.net/doc/oai-eres-id-eres2023-22 (text/html)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:arz:wpaper:eres2023_22

Access Statistics for this paper

More papers in ERES from European Real Estate Society (ERES) Contact information at EDIRC.
Bibliographic data for series maintained by Architexturez Imprints ().

 
Page updated 2025-03-30
Handle: RePEc:arz:wpaper:eres2023_22