skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: House Price Prediction via Visual Cues and Estate Attributes
The price of a house depends on many factors, such as its size, location, amenities, surrounding establishments, and the season in which the house is being sold, just to name a few of them. As a seller, it is absolutely essential to price the property competitively else it will not attract any buyers. This problem has given rise to multiple companies as well as past research works that try to enhance the predictability of property prices using relevant mathematical models and machine learning techniques. In this research, we investigate the usage of machine learning in predicting the house price based on related estate attributes and visual images. To this end, we collect a dataset of 2,000 houses across different cities in the United States. For each house, we annotate 14 estate attributes and five visual images for exterior, interior-living room, kitchen, bedroom, and bathroom. Following the dataset collection, different features are extracted from the input data. Furthermore, a multi-kernel regression approach is used to predict the house price from both visual cues and estate attributes. The extensive experiments demonstrate the superiority of the proposed method over the baselines.  more » « less
Award ID(s):
2025234
PAR ID:
10428263
Author(s) / Creator(s):
; ; ; ;
Date Published:
Journal Name:
ISVC 2022
Page Range / eLocation ID:
91=103
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Climate change-induced sea level rise (SLR) will affect a range of coastal assets and prompt difficult decisions about coastal land use across the world. Several recent studies find that current and projected SLR is associated with relatively lower property values. We contribute to this growing body of research with a case study of O‘ahu, Hawai‘i, which is famed for its beaches as well as valuable coastal real estate. We leverage a dataset that unpacks multiple types of SLR exposure and coastal parcel attributes. We apply property transaction data for the island of O‘ahu through 2019 to investigate the effect of current and expected SLR exposure on residential property prices. We find that exposed properties have already experienced declines in transaction prices, at 9 to 14%, attributed to expectations of exposure to chronic inundation (as opposed to seasonal flooding). The price declines are mainly for multi-dwelling homes as opposed to single family homes. The market response of residential properties to SLR has important implications for coastal management strategies, in particular the viability and timing of programs for retreat. 
    more » « less
  2. The availability of large datasets of organism images combined with advances in artificial intelligence (AI) has significantly enhanced the study of organisms through images, unveiling biodiversity patterns and macro-evolutionary trends. However, existing machine learning (ML)-ready organism datasets have several limitations. First, these datasets often focus on species classification only, overlooking tasks involving visual traits of organisms. Second, they lack detailed visual trait annotations, like pixel-level segmentation, that are crucial for in-depth biological studies. Third, these datasets predominantly feature organisms in their natural habitats, posing challenges for aquatic species like fish, where underwater images often suffer from poor visual clarity, obscuring critical biological traits. This gap hampers the study of aquatic biodiversity patterns which is necessary for the assessment of climate change impacts, and evolutionary research on aquatic species morphology. To address this, we introduce the Fish-Visual Trait Analysis (Fish-Vista) dataset—a large, annotated collection of about 80K fish images spanning 3000 different species, supporting several challenging and biologically relevant tasks including species classification, trait identification, and trait segmentation. These images have been curated through a sophisticated data processing pipeline applied to a cumulative set of images obtained from various museum collections. Fish-Vista ensures that visual traits of images are clearly visible, and provides fine-grained labels of various visual traits present in each image. It also offers pixel-level annotations of 9 different traits for about 7000 fish images, facilitating additional trait segmentation and localization tasks. The ultimate goal of Fish-Vista is to provide a clean, carefully curated, high-resolution dataset that can serve as a foundation for accelerating biological discoveries using advances in AI. Finally, we provide a comprehensive analysis of state-of-the-art deep learning techniques on Fish-Vista. 
    more » « less
  3. Work in computer vision and natural language processing involving images and text has been experiencing explosive growth over the past decade, with a particular boost coming from the neural network revolution. The present volume brings together five research articles from several different corners of the area: multilingual multimodal image description (Frank et al. ), multimodal machine translation (Madhyastha et al. , Frank et al. ), image caption generation (Madhyastha et al. , Tanti et al. ), visual scene understanding (Silberer et al. ), and multimodal learning of high-level attributes (Sorodoc et al. ). In this article, we touch upon all of these topics as we review work involving images and text under the three main headings of image description (Section 2), visually grounded referring expression generation (REG) and comprehension (Section 3), and visual question answering (VQA) (Section 4). 
    more » « less
  4. Images can give us insights into the contextual meanings of words, but current image-text grounding approaches require detailed annotations. Such granular annotation is rare, expensive, and unavailable in most domain-specific contexts. In contrast, unlabeled multi-image, multi-sentence documents are abundant. Can lexical grounding be learned from such documents, even though they have significant lexical and visual overlap? Working with a case study dataset of real estate listings, we demonstrate the challenge of distinguishing highly correlated grounded terms, such as “kitchen” and “bedroom”, and introduce metrics to assess this document similarity. We present a simple unsupervised clustering-based method that increases precision and recall beyond object detection and image tagging baselines when evaluated on labeled subsets of the dataset. The proposed method is particularly effective for local contextual meanings of a word, for example associating “granite” with countertops in the real estate dataset and with rocky landscapes in a Wikipedia dataset. 
    more » « less
  5. The ability to understand visual concepts and replicate and compose these concepts from images is a central goal for computer vision. Recent advances in text-to-image (T2I) models have lead to high definition and realistic image quality generation by learning from large databases of images and their descriptions. However, the evaluation of T2I models has focused on photorealism and limited qualitative measures of visual understanding. To quantify the ability of T2I models in learning and synthesizing novel visual concepts (a.k.a. personalized T2I), we introduce ConceptBed, a large-scale dataset that consists of 284 unique visual concepts, and 33K composite text prompts. Along with the dataset, we propose an evaluation metric, Concept Confidence Deviation (CCD), that uses the confidence of oracle concept classifiers to measure the alignment between concepts generated by T2I generators and concepts contained in target images. We evaluate visual concepts that are either objects, attributes, or styles, and also evaluate four dimensions of compositionality: counting, attributes, relations, and actions. Our human study shows that CCD is highly correlated with human understanding of concepts. Our results point to a trade-off between learning the concepts and preserving the compositionality which existing approaches struggle to overcome. The data, code, and interactive demo is available at: https://conceptbed.github.io/ 
    more » « less