skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Sparsity in an artificial neural network predicts beauty: Towards a model of processing-based aesthetics
Generations of scientists have pursued the goal of defining beauty. While early scientists initially focused on objective criteria of beauty (‘feature-based aesthetics’), philosophers and artists alike have since proposed that beauty arises from the interaction between the object and the individual who perceives it. The aesthetic theory of fluency formalizes this idea of interaction by proposing that beauty is determined by the efficiency of information processing in the perceiver’s brain (‘processing-based aesthetics’), and that efficient processing induces a positive aesthetic experience. The theory is supported by numerous psychological results, however, to date there is no quantitative predictive model to test it on a large scale. In this work, we propose to leverage the capacity of deep convolutional neural networks (DCNN) to model the processing of information in the brain by studying the link between beauty and neuronal sparsity, a measure of information processing efficiency. Whether analyzing pictures of faces, figurative or abstract art paintings, neuronal sparsity explains up to 28% of variance in beauty scores, and up to 47% when combined with a feature-based metric. However, we also found that sparsity is either positively or negatively correlated with beauty across the multiple layers of the DCNN. Our quantitative model stresses the importance of considering how information is processed, in addition to the content of that information, when predicting beauty, but also suggests an unexpectedly complex relationship between fluency and beauty.  more » « less
Award ID(s):
2026334
PAR ID:
10536444
Author(s) / Creator(s):
; ; ; ;
Editor(s):
Fleming, Roland W
Publisher / Repository:
PLoS computational biology
Date Published:
Journal Name:
PLOS Computational Biology
Volume:
19
Issue:
12
ISSN:
1553-7358
Page Range / eLocation ID:
e1011703
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. ABSTRACT The world around us is full of beauty. Explaining a sense of the beautiful has beguiled philosophers and artists for millennia, but scientists have also pondered beauty, most notably Darwin, who used beauty to describe sexual ornaments that he argued were the subject of female mate choice. In doing so, he ascribed a ‘sense of the beautiful’ to non‐human animals. Darwin's ideas about mate choice and beauty were not widely accepted, however. Humans may experience beauty, but assuming the same about other animals risks anthropomorphism: we might find the tail of the peacock to be beautiful, but there is no reason to believe that peahens do. Moreover, mate choice, resurrected as an object of serious study in the 1970s, simply requires attraction, not necessarily beauty. However, recent advances in psychology and cognitive neuroscience are providing a new, mechanistic framework for beauty. Here we take these findings and apply them to evolutionary biology. First, we review progress in human empirical aesthetics to provide a biological definition of beauty. Central to this definition is the discovery that merely processing information can provide hedonic reward. As such, we propose thatbeauty is the pleasure of fluent information processing, independent of the function or consummatory reward provided by the stimulus. We develop this definition in the context of three key attributes of beauty (pleasure, interaction, and disinterestedness) and the psychological distinction between ‘wanting’ and ‘liking’. Second, we show how beauty provides a new, proximate approach for studying the evolution of sexual signalling that can help us resolve some key problems, such as how mating biases evolve. We also situate beauty within a more general framework for the evolution of animal signals, suggesting that beauty may apply not only to sexual ornaments, but also to traits as diverse as aposematic signals and camouflage. Third, we outline a variety of experimental approaches to test whether animal signals are beautiful to their intended receivers, including tests of fluency and hedonic impact using behavioural and neurological approaches. 
    more » « less
  2. Abstract Neurophysiological measurements suggest that human information processing is evinced by neuronal activity. However, the quantitative relationship between the activity of a brain region and its information processing capacity remains unclear. We introduce and validate a mathematical model of the information processing capacity of a brain region in terms of neuronal activity, input storage capacity, and the arrival rate of afferent information. We applied the model to fMRI data obtained from a flanker paradigm in young and old subjects. Our analysis showed that—for a given cognitive task and subject—higher information processing capacity leads to lower neuronal activity and faster responses. Crucially, processing capacity—as estimated from fMRI data—predicted task and age-related differences in reaction times, speaking to the model’s predictive validity. This model offers a framework for modelling of brain dynamics in terms of information processing capacity, and may be exploited for studies of predictive coding and Bayes-optimal decision-making. 
    more » « less
  3. Photographer, curator, and former director of photography at the Museum of Modern Art (MoMA), John Szarkowski remarked in *William Eggleston's Guide*, "While editing directly from life, photographers have found it too difficult to see simultaneously both the blue and the sky." Szarkowski insightfully revealed a notable gap between general and aesthetic visual understanding: while the former emphasizes identifying factual elements in an image (the sky), the latter transcends mere object identification, viewing it instead as an aesthetic component--a pure expanse of blue, valued purely as a color block in visual aesthetics. Such distinctions between general visual understanding (detection, localization, etc.) and aesthetic perception (color, lighting, composition, etc.) pose a significant challenge for existing Multimodal Large Language Models (MLLMs) in comprehending image aesthetics, which is increasingly needed in real-world applications, from image recommendation and enhancement to generation. To fundamentally advance the aesthetic understanding of MLLMs, we introduce a novel dataset, PhotoCritique, derived from extensive discussions among professional photographers and enthusiasts, distinguished by its large scale, expertise, and diversity. Additionally, we propose a new model, PhotoEye, an MLLM featuring a language-guided multi-view vision fusion mechanism for understanding image aesthetics from multiple perspectives. Finally, we introduce PhotoBench, a comprehensive and professional benchmark for aesthetic visual understanding. Our model demonstrates significant advantages over both open-source and commercial models on existing benchmarks and PhotoBench. 
    more » « less
  4. Abstract—In the past decades, many graph drawing techniques have been proposed for generating aesthetically pleasing graph layouts. However, it remains a challenging task since different layout methods tend to highlight different characteristics of the graphs. Recently, studies on deep learning based graph drawing algorithm have emerged but they are often not generalizable to arbitrary graphs without re-training. In this paper, we propose a Convolutional Graph Neural Network based deep learning framework, DeepGD, which can draw arbitrary graphs once trained. It attempts to generate layouts by compromising among multiple pre-specified aesthetics considering a good graph layout usually complies with multiple aesthetics simultaneously. In order to balance the trade-off, we propose two adaptive training strategies which adjust the weight factor of each aesthetic dynamically during training. The quantitative and qualitative assessment of DeepGD demonstrates that it is capable of drawing arbitrary graphs effectively, while being flexible at accommodating different aesthetic criteria. 
    more » « less
  5. The process of capturing a well-composed photo is difficult and it takes years of experience to master. We propose a novel pipeline for an autonomous agent to automatically capture an aesthetic photograph by navigating within a local region in a scene. Instead of classical optimization over heuristics such as the rule-of-thirds, we adopt a data-driven aesthetics estimator to assess photo quality. A reinforcement learning framework is used to optimize the model with respect to the learned aesthetics metric. We train our model in simulation with indoor scenes, and we demonstrate that our system can capture aesthetic photos in both simulation and real world environments on a ground robot. To our knowledge, this is the first system that can automatically explore an environment to capture an aesthetic photo with respect to a learned aesthetic estimator. Source code is at https://github.com/HadiZayer/AutoPhoto 
    more » « less