Hostname: page-component-54dcc4c588-hp6zs Total loading time: 0 Render date: 2025-09-15T04:01:01.585Z Has data issue: false hasContentIssue false

Building Information Graphs (BIGs): remodeling building information for learning and applications

Published online by Cambridge University Press:  15 September 2025

Zijian Wang
Affiliation:
Georg Nemetschek Institute (GNI), Technical University of Munich, Munich, Germany
Rafael Sacks*
Affiliation:
Faculty of Civil and Environmental Engineering, Technion Israel Institute of Technology, Haifa, Israel
*
Corresponding author: Rafael Sacks; Email: cvsacks@technion.ac.il

Abstract

Despite significant advances in Building Information Modeling (BIM) and increased adoption, numerous challenges remain. Discipline-specific BIM software tools with file storage have unresolved interoperability issues and do not capture or express interdisciplinary design intent. This hobbles machines’ ability to process design information. The lack of suitable data representation hinders the application of machine learning and other data-centric applications in building design. We propose Building Information Graphs (BIGs) as an alternative modeling method. In BIGs, discipline-specific design models are compiled as subgraphs in which nodes and edges model objects and their relationships. Additional nodes and edges in a meta-graph link the building objects across subgraphs. Capturing both intradisciplinary and interdisciplinary relationships, BIGs provide a dimension of contextual data for capturing design intent and constraints. BIGs are designed for computation and applications. The explicit relationships enable advanced graph functionalities, such as across-domain change propagation and object-level version control. BIGs preserve multimodal design data (geometry, attributes, and topology) in a graph structure that can be embedded into high-dimensional vectors, in which learning algorithms can detect statistical patterns and support a wide range of downstream tasks, such as link prediction and graph generation. In this position article, we highlight three key challenges: encapsulating and formalizing object relationships, particularly design intent and constraints; designing graph learning techniques; and developing innovative domain applications that leverage graph structures and learning. BIGs represent a paradigm shift in design technologies that bridge artificial intelligence and building design to enable intelligent and generative design tools for architects, engineers, and contractors.

Information

Type
Position Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press

Impact Statement

Building Information Graphs (BIGs) is a new information modeling method for the architecture, engineering and construction (AEC) industry. BIGs organize and link multiple building models from different design disciplines in a graph-based format, storing them in a centralized database, representing both intra- and interdisciplinary relationships to capture design intent and constraints in a machine-understandable format. These explicit relationships enable advanced graph applications, such as change propagation and object-level version control, that are impractical with current solutions. Leveraging graph structure and graph learning, BIGs support automated, artificially intelligent, generative, and other data-centric applications for design and construction. Following computer-aided design and BIM, BIGs provide a technology foundation for the next generation of AEC software solutions.

1. Introduction

The tools used to represent and process information for building design have developed from manual drawings to descriptive geometry, and from computer-aided design (CAD) to building information modeling (BIM). Building product modeling, or BIM, emerged in the early 1990s. Eastman and Ingram separately introduced the basic concepts for digital, object-oriented representation of building information, leading the way to replacing two-dimensional (2D) drawings (Eastman, Reference Eastman1975; Ingram, Reference Ingram2020). Augenbroe (Reference Augenbroe1994a, Reference Augenbroe1994b), Björk (Reference Björk1989), Reference Björk1994) provided the first formal data structures for computer-integrated construction. The efforts of these and other researchers set in motion the R&D that led to the development of BIM tools, the Industry Foundation Classes (IFC) exchange schema, and the information management process standards that are the current state of the art (ISO19650-Reference Afsari, Eastman and Castro-Lacouture1, 2018; Sacks et al., Reference Sacks, Lee, Burdi and Bolpagni2025; buildingSMART, 2025c).

However, despite advances in BIM technologies and processes, multidisciplinary design teams in the architecture, engineering and construction industry (AEC) still face numerous challenges when generating, sharing, and using building information. The challenges include bulk exchange of information for coordinating models across the boundaries of professional design disciplines as designs are iteratively developed; maintaining consistency across different versions of federated building model files; design review and model checking for conformance to requirements; preprocessing of models needed before the application of sophisticated tools for generative design and for performance simulations and analyses; and managing and leveraging information through the life of a project beyond the design phase (Borrmann et al., Reference Borrmann, König, Koch, Beetz, Borrmann, König, Koch and Beetz2018; de Wilde, Reference de Wilde2019; Bloch, Reference Bloch2022; Wang et al., Reference Wang, Ouyang and Sacks2023).

A key barrier is that state-of-the-art BIM authoring tools restrict sharing of design information to a building’s geometry and alphanumeric properties, largely excluding information that defines design intent and constraints. In the AEC domain, design intent refers to how designers envision buildings, spaces, and building elements to be configured so that they can fulfill their functionality, constructability, and performance criteria. The explicit expression of design intent is primarily concerned with ensuring that the objects are reconfigured automatically by software when design changes are made by operators. Parametric constraints on geometry, on materials, and on topological relationships between a building’s elements are a mechanism for explicitly expressing design intent (Sacks et al., Reference Sacks, Eastman and Lee2004; Ji et al., Reference Ji, Borrmann, Beetz and Obergrießer2013).

A key difficulty in the AEC industry arises from the diverse nature of the expertise of the numerous professional disciplines needed to design a building. Architects, civil and structural engineers, mechanical engineers, electrical engineers, ventilation, heating, hydraulics, lighting, and acoustics consultants, to name but a few, conceive of buildings and their constituent parts in quite different ways. To an architect, drywall partitions divide spaces; to a structural engineer, they do not exist; to a construction planner, they require modeling in at least three different basic constituent parts for work planning. Therefore, designers require software that models building systems with distinctive object classes, relationships, and behaviors that are useful for their specific discipline. This makes it difficult for any discipline-specific design software to capture and express design intent and constraints that transcend the discipline.

Another difficulty is that the exchange schema must reduce to the lowest common denominator, and is therefore based on universal concepts, with the exchange files containing only what can be sensibly shared. Existing information exchange schema, such as IFC (ISO 16739-Reference Afsari, Eastman and Castro-Lacouture1, 2018), were not designed to exchange parametric objects, and they can neither capture nor express design intent concepts and the explicit constraint relationships required to embody them (Borrmann et al., Reference Borrmann, Hyvärinen and Rank2009). With very few exceptions (such as the constraint that windows and doors must be embedded in walls), discipline-specific relationships that express design intent (e.g., reinforcing bars must be wholly contained within concrete elements, pipelines must have connectivity along their run) are not embedded in exchange files. Furthermore, design intent relationships that express logic across disciplines—for example, pipes that must be attached to and supported by structural elements, concrete columns that must be wholly contained in architectural representations of the same columns, and so on—find no explicit expression in any of the data. Where any of these relationships are required for some design or analysis function, they must be assumed or added explicitly by users. Where possible, they can be inferred in a process known as semantic enrichment (Belsky et al., Reference Belsky, Sacks and Brilakis2016; Bloch and Sacks, Reference Bloch and Sacks2018).

The Linked Building Data (LBD) research community has explored the use of graphs to represent and integrate built environment data. Early research investigated semantic web techniques to structure building information as knowledge graphs, aiming to enhance collaboration within the AEC industry (Beetz, Reference Beetz2009). Researchers envision seamless linkage of all project data in standardized forms, enabling interoperable software communication and efficient data flow (Pauwels et al., Reference Pauwels, Costin and Rasmussen2022). To achieve this, the LBD community adopted techniques from the semantic web for compiling and querying graphs, including Web Ontology Language (OWL) for constructing ontologies, Resource Description Framework (RDF) for instance graphs, and SPARQL for graph queries (Pauwels, Reference Pauwels2014). Additionally, domain-specific ontologies have been developed, such as ifcOWL (derived from the IFC schema) (Beetz, Reference Beetz2009; Pauwels and Terkaj, Reference Pauwels and Terkaj2016) and the Building Topology Ontology (BOT) for capturing building topology (Rasmussen et al., Reference Rasmussen, Lefrançois, Schneider, Pauwels and Janowicz2020). However, current ontologies predominantly focus on individual BIM disciplines or integration with external data, and there is a lack of definition of interdisciplinary object relationships.

Leveraging semantic web techniques and predefined ontologies, researchers have developed tools such as IFCtoLBD to convert IFC STEP files into instance knowledge graphs represented in RDF format (Bonduel et al., Reference Bonduel, Oraskari, Pauwels, Vergauwen and Klein2018). This approach achieves the goal of compiling BIM design models as BIM knowledge graphs, although many object attributes and all geometries are omitted in the IFCtoLBD converter, resulting in incomplete model information. Under their application scenarios, some attributes were considered unimportant or irrelevant, and geometry was too complicated to be handled, which limited the downstream applications (Bonduel et al., Reference Bonduel, Oraskari, Pauwels, Vergauwen and Klein2018). It is important to note that the primary goal of LBD knowledge graphs is to integrate heterogeneous built environment data into a uniform format that satisfies interoperability and retrieval requirements (Pauwels et al., Reference Pauwels, Costin and Rasmussen2022; LBD, 2024). The knowledge graphs do not address the representation of complex and complete BIM design data to support machine learning.

Instead, progress in artificial intelligence (AI) offers great promise that is yet unrealized in design. A common thread across successful applications of AI is that the data are modeled and stored in suitable formats. Algorithms and learning strategies are tailored to the data representations. However, up to now, researchers and developers in the AEC industry have simply adopted tools developed in computer science and applied them to AEC challenges without considering the inherent richness of digital building information. Design models encompass multimodal data, ranging from three-dimensional (3D) vector geometry to attributes, and from text to relational data. The richness of modality makes it difficult to employ machine learning directly. For example, object classification is done by generating 2D images from 3D spatial BIM geometries as input for vision-based deep learning models, which fails to utilize much of the available building information (Koo et al., Reference Koo, Jung and Yu2021). Similarly, object geometries are sampled as point clouds or graphs for 3D deep learning techniques (Koo et al., Reference Koo, La, Cho and Yu2019; Collins et al., Reference Collins, Braun, Ringsquandl, Hall and Borrmann2021). This approach ignores finding a suitable way to structure the complex and multimodal building information, leading to narrow and suboptimal applications.

In light of these limitations and opportunities, we conclude that the current state-of-the-art BIM information representations are inadequate. A better way to express information for building design and construction is needed. We propose Building Information Graphs (BIGs) as an alternative information modeling method to structure design information in a graph-based format that can express and store design intent and constraints in a central database. In BIGs, discipline-specific design models are compiled as subgraphs in which nodes model building objects and edges model building object relationships. Additional nodes and edges in a meta-graph link the building objects across subgraphs. Graphs offer a flexible and machine-accessible format to integrate multimodal design data that can be processed and embedded for learning. Consequently, diverse applications can be developed to leverage graph structures and graph learning, thereby supporting and automating design and multidisciplinary collaboration.

The next section defines BIGs, lists their essential properties, and lays out three key challenges that researchers will need to confront. The three subsequent sections discuss the considerations around these challenges:

  • Section 3 addresses the questions concerning the concepts that will be necessary for the relationship class ontology.

  • Section 4 considers how graph learning can be developed for many graph-based tasks, such as link prediction, node classification, and graph generation.

  • Section 5 provides a comprehensive summary of graph-based applications that leverage object relationships, achieving functionalities beyond the scope of current BIM technologies.

Each of these sections concludes with a set of specific research questions. The concluding section summarizes the key points and calls for concerted research and development to explore the potential of BIGs to be the next paradigm change for information representation in the AEC industry.

2. Building information graphs

2.1. Definition

BIGs represent design information in a graph-based data format. The definition of a BIG $ G $ is as follows:

(2.1) $$ G=\left(\underset{i=1}{\overset{n}{\cup }}\;{G}_i\right)\cup {G}_{meta} $$

where:

- $ G $ is the linked graph representation of building models.

- $ {G}_i $ represents the subgraph of a building discipline $ i $ , such as architecture, structural engineering, or mechanical engineering.

- $ {G}_{meta} $ represents the meta graph that contains interdisciplinary relationships to link subgraphs.

(2.2) $$ {G}_i=\left({V}_i,{E}_i\right) $$

- $ {V}_i $ is the set of vertices corresponding to objects within a building discipline $ i $ .

- $ {E}_i $ is the set of edges representing relationships between the objects within the discipline $ i $ (intradisciplinary relationships).

(2.3) $$ {G}_{meta}=\left({V}_{meta},{E}_{meta}\right) $$

- $ {V}_{meta} $ is the set of vertices corresponding to objectified interdisciplinary relationships.

- $ {E}_{meta} $ is the set of edges representing interdisciplinary relationships.

In mathematics, graphs consist of a set of vertices and edges (Bondy and Murty, Reference Bondy and Murty2008; Bollobás, Reference Bollobás2013). BIGs are defined as a collection of discipline-specific subgraphs ( $ {G}_i $ ), such as architectural and structural subgraphs, which are linked through a meta-graph ( $ {G}_{meta} $ ) that encapsulates interdisciplinary relationships, as shown in Figure 1. In the subgraphs, vertices (nodes) represent objects, including building elements such as doors, columns, and spaces. Edges (links) define intradisciplinary relationships, representing associations within a single domain (e.g., adjacency between two walls in the architectural subgraph). In the meta-graph, nodes represent functions, such as versions and constraints, and edges define interdisciplinary relationships, linking objects across disciplines (e.g., a hosting relationship between an architectural wall and a structural column). Interdisciplinary relationships in the meta-graph can also be objectified as nodes with additional attributes. For example, a constraint edge between a wall and a column can be represented by a node in the meta-graph with additional attributes that specify the condition, operator, and status.

Figure 1. Illustration of Building Information Graphs ( $ G $ ) in partial models.

There are two key differences between BIGs and BIM models:

  1. 1. BIGs model represent and store interdisciplinary relationships. BIM models focus on discipline-specific objects, attributes, geometries, and limited intradisciplinary relationships. In BIGs, both intradisciplinary and interdisciplinary relationships are considered. For example, when an expert observes the models from three different disciplines shown in Figure 1, the expert understands that the radiator (a mechanical/electrical/plumbing [MEP] object) is intentionally attached to the wall (an architectural object) and intentionally placed below the window. However, in BIM models, the attached to information and the relative placing information are implicit—no current method or technology enables machines to understand the interdisciplinary relationships nor to store the semantics.

  2. 2. In a BIG, the information describing a building is stored in a graph-based database rather than in files. Existing BIM technologies use files to store and exchange design information. In BIGs, the information from all disciplines is stored using the same graph database technology with subgraphs for each discipline and a meta-graph that implements relationships between the objects from different subgraphs.

The advantages of BIGs are twofold. First, BIGs support graph machine learning. A design model contains rich, complex, and multimodal data—including attributes, relationships, geometries, and more—yet existing BIM tools and exchange formats are not structured to facilitate machine learning. By leveraging a flexible and learning-suitable representation, BIGs integrate these diverse data and process them into vector spaces for learning, aiming to bridge building design and AI.

Second, BIGs enable more automated and intelligent design applications. Representing relationships explicitly introduces additional contextual information, making sophisticated functionalities feasible, such as multidisciplinary change propagation and object-level version control, which are difficult to implement in current BIM systems.

In summary, BIGs adopt graphs to represent and store object relationships both within and across design disciplines, and the design data are stored in a graph-based database rather than traditional files. Additionally, BIGs integrate and embed rich, complex, and multimodal BIM data into a learning-suitable representation for supporting AI. By leveraging the graph structure and graph learning, more automated and intelligent functionalities in design and construction become feasible.

2.2. Functional properties

To represent the richness of information in building models, and to support the full range of possible computations and design applications that are needed, BIGs will need to exhibit the following functional properties:

Attributes on nodes, edges, and graphs: Nodes, edges, and graphs can all carry attributes. Node attributes describe the semantics of building objects, such as class, Globally Unique Identifier (GUID), width, length, and material. Edge attributes describe the properties of relationships between objects (e.g., adjacency or constraint violation). Graph-level attributes capture metadata, such as the discipline of subgraphs (e.g., architecture or structure) or version identifiers.

Directional edges: BIGs require both directed and undirected edges. Directed edges represent dependencies (e.g., a column supporting a slab) and are used for causal or hierarchical relationships. Undirected edges capture mutual relationships (e.g., adjacent walls). Depending on the context, BIGs can be treated as directed, undirected, or mixed graphs.

Dynamic definition: BIGs are developed dynamically over time, with nodes, edges, and subgraphs being added, removed, or modified to reflect design changes. Nodes and edges, and even whole subgraphs that are superseded with new information (new nodes and edges) are not deprecated, but rather their metadata are changed to reflect their status. This feature can support functionalities such as incremental updates, version control, undo steps to revert to earlier versions, and incremental increase of the Level of Development (LOD) of objects, as corresponding objects of different detail can be associated with one another to support storage of a record of the design steps.

Aggregation: Object nodes in BIGs can be related through aggregation. Based on the nature of the subordinate relationships, aggregation can be divided into composition/decomposition and association. For example, a structural frame can be decomposed into beams and columns, and a detailed architectural wall can be decomposed into a structural core and two finishing layers. Association denotes a relationship between objects that are not parts of the parent object, but simply associated with it, such as furniture belonging to a room.

Layering: Data in BIGs can be organized into layers to support structured data management. For example, as an architectural model evolves from conceptual to detailed design, graphs from the two stages can form separate layers, with corresponding objects, such as a general wall and its detailed version, linked to represent the design progression. Layers are not limited to graphs. As noted in Section 5.6, life cycle data management places the graph above the file storage layer, using it to link files across project phases. Layering in BIGs primarily serves data organization and management.

Multimodality: Design models involve diverse information types and data modalities. Graphs compile building objects, semantics, and relationships, but other data (e.g., geometry, sensors, and documents) may need to remain in their native formats. BIG supports linking non-graph data to the graph for seamless integration. For example, the geometry of a wall is stored in a geometry file, and the file address is inserted in the corresponding node as a pointer attribute. Therefore, a BIG acts as a central repository for data integration.

Computability: Information in graphs can conveniently be converted into different formats that facilitate computation, such as embedding multimodal data into vector spaces for statistical learning or IF-THEN statements for rule-based inferencing.

2.3. Three key challenges

Representing relationships among building objects explicitly introduces a new dimension of data that captures topology, design intent, and constraints among building objects. Leveraging object relationships, a range of innovative graph applications that are hard to achieve using current BIM technologies become possible, such as multidisciplinary change propagation, automated consistency maintenance, object-based version control, and more. Graphs also provide a machine-accessible format to facilitate learning. However, this is an emerging domain, and we identify three key challenges:

  1. 1. What new classes of object relationships will be needed? How can meta-graph relationships be instantiated automatically, accurately, and efficiently?

  2. 2. What graph learning techniques are needed to effectively apply AI methods to building design?

  3. 3. What applications could be developed by leveraging graphs if BIGs become available?

It is noteworthy that in the following discussion, our goal is to raise the key challenges and elaborate on the questions, but not to provide complete answers. Instead, we review and summarize existing work in the AEC field and in other domains, including computer science, to explore whether their experience and knowledge can be leveraged to tackle the challenges of information representation for AEC.

3. Object relationships

Numerous schemas and ontologies have been developed for information exchange in the AEC industry in general and for BIM models in particular. Across these, the hierarchies for classifying physical building objects and spaces are well developed, whereas the hierarchies of classes that represent the possible relationships between these objects are limited. Furthermore, none of the schemas define the relationships needed for cross-disciplinary design coordination and collaboration. The IFC v4.3 ADD2 schema, for example, has 876 class definitions (buildingSMART, 2025c). Of these, 99 classes represent parts of buildings and the relationships between them. Six classes define spaces, and 88 classes are physical building entities (45 shared building elements, 26 building service elements, 10 components, and seven facility elements), and most of these have multiple enumerated types. The remaining five are entities that objectify relationships between physical elements and/or spatial elements. These include aggregation (groupings of objects), composition and decomposition (assemblies of component parts), structural or physical connection (e.g., beams to columns or duct or pipe elements to one another in service lines), spatial containment (i.e., entities within spaces), and positioning (placement of an element relative to others). However, we identify at least four additional types of relationships that can exist between building elements: correspondence, equivalence, function, and spatial topology. These relationships are essential for complete semantic representation, and thus for effective and efficient application of intelligent software applications.

Correspondence between two building objects from two different BIM models denotes that they represent the same object in the real world. For example, an architect may model a column in a building as a single object that rises through all of a building’s stories. A structural engineer, on the other hand, would likely model a separate column on each floor, so that each individual instance could be assigned a different cross-section and reinforcement arrangement. Each structural column corresponds to the full height architectural column. Should the latter be deleted or moved, the former should be changed accordingly to maintain the consistency of the design.Footnote 1

Equivalence exists when two objects in different discipline models represent a single physical object. For example, a wall or a floor in an architectural BIM model may contain a concrete structural layer. In the structural BIM model, the same object is modeled simply as a structural wall or a slab. The concrete core of a column in an architectural model and the corresponding structural concrete column represent the same object, and they are equivalent in every sense (material definition, geometry, etc.).

Functional intent relationships express the way in which an object from one BIM model functions with respect to an object from the BIM model of a different discipline. In general, this means that one object performs a function that is necessary for the other object (this type can also be called a design intent relationship). A sprinkler for fire control placed on a wall directly above a stove in a residential kitchen is a good example of a functional relationship that reflects the intent of the designers that the sprinkler should be suitably positioned to be able to extinguish a fire on the stove. The sprinkler is commonly modeled in the MEP BIM model, whereas the stove will appear in the architectural BIM model. A drain pipe mounted on a wall is another example—in this case, the drain pipe is supported vertically by brackets that affix it at some given distance to the architectural wall. A hole in a structural concrete wall through which an MEP ventilation duct passes is a third example.

A functional intent relationship requires a property denoting its intended function, or purpose, and a property defining the nature of the intended topological constraint (e.g., the sprinkler must be above the stove and aligned with the X axis of the center of the stove; the pipe must be vertically parallel to the outer face of the wall some fixed offset distance from it). The properties are accommodated by objectifying the relationship in the meta-graph with a node. The constraints indicate how either object must be repositioned in response to a change in the position of the other object made by the designer of the latter object, in order to maintain the integrity of the functional design intent.

A spatial topology relationship exists implicitly between any and every pair of building elements in a federated BIM model, but it needs only to be instantiated between the elements’ nodes in a BIG where the relationship is needed for a correspondence or a functional intent relationship. A topological relationship defines the existing logical relative location of any two objects. Among the possible spatial topology relationship values between two objects in each of the X/Y/Z directions are CONTAINS/CONTAINED-IN, ALIGNED, EXACT_OVERLAP, LEFT/FRONT/ABOVE, LEFT/FRONT/ABOVE_CONTACT, LEFT/FRONT/ABOVE_OVERLAP, RIGHT/BACK/BELOW, RIGHT/BACK/BELOW_CONTACT, and RIGHT/BACK/BELOW_OVERLAP. The values can be computed using exact or bounding box geometry. Bounding box topology relationships are fast to compute and can provide sufficient information in many cases (Daum and Borrmann, Reference Daum and Borrmann2014); exact geometry relationships need only be computed when the bounding box relationships are not unequivocal (Ouyang, Reference Ouyang2023).

What distinguishes these relationships from the relationships defined in existing schemata is that they can express the logic of design intent that crosses the boundaries between different disciplines. The IFC schema was developed on the basis of necessity—the need to export discipline-specific BIM models to files that contain the information encapsulated in the native schema of the exporting application. No new information is generated during the export, and, therefore, no relationships with the information from other design disciplines need to be included—thus, none of these relationship classes were relevant for the IFC schema. However, this information is essential for any software application to maintain consistency of a holistic building representation that is composed of federated models from different disciplines.

In addition to these four, two of the relationships defined in IFC (aggregation and spatial containment) are also relevant for establishing the meta-graph relationships between nodes that represent building elements that belong to different discipline subgraphs. The IFC relationship connection can serve as one of a group of relationships that express functional intent. The last two (composition and decomposition and positioning) are relevant only within discipline subgraphs.

Research conducted in the framework of the European Union funded Cloud BIM (CBIM) project has shown that checking for and proactively maintaining consistency in building models stored as subgraphs is feasible when correspondence, function, and spatial topology relationships are present (Wang et al., Reference Wang, Ouyang and Sacks2023) (noting, of course, that automated edits require explicit permission from the owners of the affected building objects). Ouyang et al. (Reference Ouyang, Wang and Sacks2023) also showed that some of these relationships can be supplemented automatically. However, despite the promising results of early research, many questions remain to be answered in the R&D of BIG systems. Among them:

  1. Q1 Considering the use-case of checking and maintaining logical and spatial consistency of design information across the different discipline models in a BIG, what aspects of functional multidisciplinary design intent are required? What types of constraints are needed to define the function relationships? What are the necessary geometry, topology, and positioning parameters of the constraints?

  2. Q2 Can we define types of functional design intent relationships and their constraints, and classify them in an ontological framework? What should be the scope and granularity of such an ontology, to balance usability and computational efficiency? Can graph learning methods be applied to identify and “learn” types of functional design intent relationships?

  3. Q3 What AI algorithms can be used to semantically enrich BIGs with interdisciplinary relationships? In other words, what graph learning techniques can efficiently and accurately instantiate interdisciplinary, multiclass relationships to form the meta-graph? To what extent can their instantiation be automated?

  4. Q4 From the practical standpoint of graph size, complexity, storage, and manipulation, should the different relationships be computed and instantiated as the building elements are appended to or edited in the subgraphs, or should they be computed and instantiated only when needed (in response to a user request to check and maintain integrity?

  5. Q5 Where corrections are needed to a design, what mechanisms are needed to change the nodes and edges that represent the geometry, location, properties, behavior, and intra- and interdiscipline relationships to achieve conformance with all constraints? Are they feasible and tractable?

4. Graph learning

In this section, we explore the methods and challenges of applying graph learning to BIGs. We first review the achievements of computer vision (CV) and natural language processing (NLP), identifying key lessons that can enlighten our approach. These include (1) fundamental data representation formats, (2) learning techniques tailored to those formats, (3) specialized training methods, and (4) the role of large-scale datasets (Section 4.1). Inspired by these insights, we propose the graph learning framework for BIGs in Section 4.2, from graph representation to embedding, to learning, and to training. Delving into the proposed framework, we explain and discuss the associated challenges and potential solutions for each of these aspects (Sections 4.34.5). We emphasize that the goals of graph learning are to support domain-specific applications (Section 4.6), and to support the generation of the meta graph. Lastly, we summarize the key open challenges that remain to be addressed in Section 4.7.

4.1. Experience from AI domains

Reflecting on the evolution of AI, particularly in the domains of CV and NLP, we note that each domain has its own purpose-specific representation of information for learning, suitable algorithms and training methods, and publicly available datasets. However, as illustrated in Figure 2, the BIM domain remains unexplored. This section discusses these four aspects (data representation, algorithms, training methods, and datasets) in the development trajectories of CV and NLP.

Figure 2. Data representation in various domains (this figure is modified based on Wang et al. (Reference Wang, Ying and Sacks2024)).

When converting human-understandable raw data into machine-accessible and learning-suitable representations, different AI domains adopt distinct approaches. In a CV, images are stored as pixel arrays representing color information. These arrays are structured as matrices that capture the spatial locality of images and can be efficiently manipulated using matrix operations. In NLP, natural language is processed as tokens and mapped to numerical representations. Tokenization involves splitting text into smaller units, such as words or subwords, and mapping each token to a unique integer based on a vocabulary. For example, Figure 2 illustrates an example of word-based tokenization, where each word stands for a token and each token is converted into a unique integer. During this process, tokens preserve the sequential nature of language, allowing the downstream algorithms to capture semantic and syntactic dependencies efficiently.

Algorithms are often specifically designed to leverage the unique features of the data representation. In CV, Convolutional Neural Networks (CNNs) were developed to extract spatial features, such as edges, textures, and shapes, from image matrices (LeCun et al., Reference LeCun, Bottou, Bengio and Haffner1998). CNNs use convolutional operations, where kernel filters slide across pixel neighborhoods to detect patterns, and pooling layers downsample spatial information for efficient feature extraction. These operations are tailored to the matrix representation.

In NLP, algorithms have also been designed to address the challenges of representing languages in sequenced tokens. Word2Vec, a seminal embedding work, maps tokens to dense, high-dimensional vectors, capturing semantic and syntactic relationships between words (Mikolov, Reference Mikolov2013). These embeddings allow machines to process language in a meaningful and efficient way. More recently, the Transformer architecture has profoundly impacted NLP by introducing the attention mechanism, which allows the model to weigh the importance of different words in a sentence relative to each other (Vaswani, Reference Vaswani2017). Because the token representation lacks inherent sequential structure during parallel processing, transformers incorporate positional encoding to enable the attention mechanism to capture sequential dependencies while processing tokens in parallel (Vaswani, Reference Vaswani2017).

The success of generative pretrained transformers (GPT) underscores the critical importance of training methods tailored to a domain’s scenarios. ChatGPT, released in 2022 based on the GPT-3.5 series models, marked the first time a large language model captured widespread public attention (OpenAI, 2022b). The dataset and algorithm architecture of GPT-3.5 and GPT-3 remained largely unchanged (OpenAI, 2022a). However, GPT-3.5 adopted Reinforcement Learning from Human Feedback (RLHF) as the fine-tuning method (OpenAI, 2022a). RLHF aligns the model’s outputs with human preferences and values by incorporating human feedback into the training loop. Specifically, human labelers prepare desired outputs for the language model in the stage of supervised learning. Following that, human labelers rank the quality of the language model’s responses to train a reward model, which is used to optimize the language model for aligning human preferences in reinforcement learning. Again, with similar algorithm architecture and datasets, this innovative training method enabled GPT-3.5 to generate more relevant and user-aligned responses, setting it apart from prior GPT models.

OpenAI-O1, which uses the strategy of the chain of thought to resolve complex problems step by step, outperforms previous GPT-series models (OpenAI, 2024). Recently, the DeepSeek-R1 model has been shown to perform comparably to OpenAI-O1, while requiring less computation for training (DeepSeek-AI et al., Reference Guo, Yang, Zhang, Song, Zhang, Xu, Zhu, Ma, Wang, Bi, Zhang, Yu, Wu, Wu, Gou, Shao, Li, Gao and Zhang2025). One strategy adopted by DeepSeek is to train a smaller student model distilled from a larger teacher model while retaining similar performance, a process known as distillation (Hinton et al., Reference Hinton, Vinyals and Dean2015). This demonstrates the effectiveness and economic efficiency of distillation on powerful base models and larger-scale reinforcement learning (DeepSeek-AI et al., Reference Guo, Yang, Zhang, Song, Zhang, Xu, Zhu, Ma, Wang, Bi, Zhang, Yu, Wu, Wu, Gou, Shao, Li, Gao and Zhang2025). It further highlights that advancements in training techniques and engineering optimizations can enhance model performance alongside improvements in algorithms and computational resources.

Additionally, a critical factor in advancing AI is the construction of large-scale datasets. In CV, ImageNet was the first large-scale image dataset, initially containing 3.2 million images across 5247 classes in 2009 (Deng et al., Reference Deng, Dong, Socher, Li, Li and Fei-Fei2009). By 2014, ImageNet had expanded to over 14 million images spanning 20,000 classes (Russakovsky et al., Reference Russakovsky, Deng, Su, Krause, Satheesh, Ma, Huang, Karpathy, Khosla, Bernstein and Fei-Fei2015). To maximize its potential, ImageNet was used for the annual ImageNet Large Scale Visual Recognition Challenge (ILSVRC), starting in 2010. In 2015, ResNet, a deep learning model, achieved a top-5 error rate of 3.57% in the ILSVRC, outperforming the estimated human error rate of ~5% for the same task (He et al., Reference He, Zhang, Ren and Sun2016). This marked the first time deep learning surpassed human performance in image object classification, demonstrating that large, high-quality datasets, combined with advanced algorithms and sufficient computational power, enable machines to outperform humans in specific tasks. Similarly, GPT-3.5 was trained on a diverse mix of datasets, including Common Crawl, which contains over 250 billion web pages before filtering and cleaning (Crawl, Reference Crawl2024).

In summary, we have identified four aspects of AI development:

  1. 1. A learning-suitable data format that can represent the unique features of the domain information is essential. For instance, natural language is tokenized into sequences that preserve word order, enabling models to capture semantic and syntactic relationships.

  2. 2. Algorithms are designed to leverage the specific properties of the data. In CV, operations such as convolution and pooling in CNNs are optimized to extract spatial features from image matrices.

  3. 3. Training methods are tailored to domain requirements. For example, RLHF fine-tunes language models to align their outputs with human preferences by incorporating human feedback in a training loop. Distillation is used to retain the comparable performance of LLMs while reducing computation resources.

  4. 4. The construction of large-scale datasets is fundamental to unleashing the potential of AI. Datasets like ImageNet in CV and Common Crawl in NLP provide the foundation for training models that excel in their respective tasks.

While it is important to acknowledge the critical role of hardware advancements (such as Graphics Processing Units) to enable the training of complex models, the development of computational hardware is beyond the scope of this study.

4.2. Graph learning for BIGs

Inspired by the experience from successful AI domains, we rethink how AI can be leveraged for BIM design. Here, we propose a conceptual framework of graph learning for building information models, as shown in Figure 3.

Figure 3. Conceptual graph learning for building information models.

First, graphs are adopted to compile and represent design data, as their flexibility enables effective integration of heterogeneous data, and graphs provide a format suitable for learning (Section 4.3). Second, the learning process is divided into two stages: embedding and pattern recognition. Embedding transforms multimodal BIM data—including geometry, attributes, and relationships—into a high-dimensional vector space, enabling efficient computational processing. Subsequently, pattern recognition algorithms leverage the resulting embeddings to detect statistical patterns and accomplish various graph-based tasks (Section 4.4). Training strategies should be specifically tailored to domain characteristics, supported by the construction of comprehensive, large-scale datasets (Section 4.5). The ultimate purpose of these learning processes is to support design-related applications, such as data enrichment and generative design (Section 4.6).

4.3. Graph representation

Design data are heterogeneous and complex. A BIM design model encompasses various types of data, including object geometries that define shapes and locations, attributes and properties for object behaviors, and relationships (Sacks et al., Reference Sacks, Lee, Burdi and Bolpagni2025). Additionally, a project includes user requirements and codes in PDF, 2D drawings, tabular schedules, and so on. Beyond these, BIM authoring software records behavioral data, such as user operation histories during the modeling process.

For collaboration and communication purposes, design models are typically exported from software into IFC files in the STEP format (IFC-SPF), where IFC serves as a neutral and open data schema (Borrmann et al., Reference Borrmann, König, Koch, Beetz, Borrmann, König, Koch and Beetz2018; buildingSMART, 2025b). During the export process, part of the design information is lost due to the unmatched data schemas (Inhan Kim and Seo, Reference Inhan Kim and Seo2012), and translation errors may occur (Venugopal et al., Reference Venugopal, Eastman and Teizer2015). From an interoperability perspective, IFC acts as a bridge for exchanging design models, enabling machines to access and allowing users to view models (Borrmann et al., Reference Borrmann, König, Koch, Beetz, Borrmann, König, Koch and Beetz2018).

Nevertheless, IFC was not originally designed for machine learning purposes. Even for simple projects, IFC-SPF tends to be complex and cumbersome. For example, the spatial location of a wall in an IFC-SPF file is stored using “IfcLocalPlacement,” which iteratively defines the relationship between the wall and other objects until it eventually references a specific point “IfcAxis2Placement3D.” While this pointer-based approach avoids redundant storage of the same data, the long and implicit chains make it challenging for machines to learn patterns.

The buildingSMART community has developed a series of tools to parse IFC-SPF files into other formats for wider usage (buildingSMART, 2025b), such as JavaScript Object Notation formats intended primarily for web-based data exchange (Afsari et al., Reference Afsari, Eastman and Castro-Lacouture2017) and RDF graphs for data integration and query (Pauwels and Terkaj, Reference Pauwels and Terkaj2016; Pauwels, Reference Pauwels2021). Users can read and use the resulting files more easily than they can IFC data files, despite the fact that they still conform to the complex data structures of the IFC schema. For example, building elements are still located by chains of “IfcLocalPlacement” objects that must be followed to derive global placement coordinates.

To simplify the graph representation, the LBD community has developed ontologies to guide the compilation of knowledge graphs in a more efficient manner (LBD, 2024). For example, the BOT ontology offers a direct structure to describe building topology (Rasmussen et al., Reference Rasmussen, Lefrançois, Schneider, Pauwels and Janowicz2020). However, during the conversion process, only partial design information is kept, while many attributes and all geometries are discarded (Bonduel et al., Reference Bonduel, Oraskari, Pauwels, Vergauwen and Klein2018). This limits the potential of machines to understand the design model as a human expert might, as the important 3D geometry information is omitted.

Although IFC models in various digital formats are technically machine-accessible, the inherent complexity of their data schema limits ease of implementation and practical use. Recognizing these issues, buildingSMART aims to revise the existing IFC schema to enhance compatibility with AI in the future (buildingSMART, 2024). We argue that the purpose of machine learning in the BIM domain should be oriented toward developing design applications, such as generative design and collaboration tools, rather than addressing the complexity of data schemas.

Therefore, we combine several existing ontologies and adopt a graph-based format to concisely represent and integrate design data in a way that could support machine learning. In BIGs, design data are organized using a core-extension layer structure: information suitable for graph representation is compiled into a core graph, whereas data not efficiently represented as graphs are retained in their native file formats and linked to the core graph (Pauwels et al., Reference Pauwels, Costin and Rasmussen2022; Ouyang et al., Reference Ouyang, Wang and Sacks2023). Specifically, in the core graph layer, nodes correspond to objects, node features encode object attributes, and edges represent relationships among objects. Other data are stored in their native formats, such as geometries in Polygon File Format (PLY) files. The file addresses are stored as node attributes in the core graph. This can effectively integrate the multimodality of design data and ensure efficient retrieval. A conceptual illustration of this representation is provided in Figure 3 (Representation).

The core-extension structure provides a centralized and graph-based representation to integrate diverse building data. A fixed and uniform schema for this integration is essential for achieving interoperability and consistency across software databases. We leverage existing ontologies to represent data in a customized and concise manner. In the subgraph, the BOT ontology is used to represent the building topology (Rasmussen et al., Reference Rasmussen, Lefrançois, Schneider, Pauwels and Janowicz2020). The terminology of object types and properties follows the buildingSMART Data Dictionary (bSDD), which is a collection of definitions of terms to describe the built environment (buildingSMART, 2025a). To link geometry files in the extension layer, the File Ontology for Geometry Formats (FOG) is adopted (Bonduel et al., Reference Bonduel, Wagner and Pauwels2020). Beyond subgraphs, the initial version of CBIM ontology can be used to represent interdisciplinary relationships, such as correspondence or equivalence (Ouyang, Reference Ouyang2023).

Lastly, we distinguish graph-based data storage from graph representation for machine learning. Depending on the application, the graph may require reorganization to align with specific learning objectives. For instance, in tasks such as space-type classification, a tailored graph might consist of space nodes and spatial relationships. In this study, we explore graph learning techniques that can be applied broadly to BIGs. The input is a multimodal design model, and the output could be high-dimensional vectors that are organized into a graph structure, facilitating generic and diverse domain tasks downstream.

4.4. Embedding and learning

Graphs can represent and integrate building information from software and files. Following that, embedding aims to process multimodal design data, such as geometries, attributes in texts, and relationships in graphs, to a high-dimensional vector space, from human-understandable data to a learning-suitable representation. Then, algorithms can leverage these embeddings to recognize statistical patterns serving for downstream applications. The conceptual learning process, integrating both embedding and pattern recognition, is shown in Figure 3.

4.4.1. Embedding multimodal design data

Unlike embedding for NLP, where a single information type (text) is embedded in dense high-dimensional vectors to capture semantic and syntactic relationships between words, building design data embedding is challenging because design models containing at least geometries, relationships, and attributes are heterogeneous.

In recent years, multimodal large language models (MLLMs) have advanced the integration of multiple data types, particularly vision and language (Lin et al., Reference Lin, Lee, Shoeybi, Lin, Catanzaro and Ping2024; Liu et al., Reference Liu, Wang, Shao, Luo, Qiao, Shou, Zhang and You2024). These models utilize modality-specific encoders, such as CNNs for images and transformers for text, to generate embeddings tailored to each modality. These embeddings are then mapped into a shared vector space through a process called alignment, facilitating unified learning across modalities.

For example, CLIP (Contrastive Language–Image Pretraining) aligns text and image data into a shared embedding space using modality-specific encoders and contrastive learning (Radford et al., Reference Radford, Kim, Hallacy, Ramesh, Goh, Agarwal, Sastry, Askell, Mishkin, Clark, Krueger and Sutskever2021). By associating textual descriptions with corresponding visual content, CLIP enables seamless interaction between modalities using an alignment architecture. Its success highlights the importance of embedding heterogeneous data into a unified and shared vector space (Radford et al., Reference Radford, Kim, Hallacy, Ramesh, Goh, Agarwal, Sastry, Askell, Mishkin, Clark, Krueger and Sutskever2021).

Studies in MLLMs provide valuable inspiration for researchers of AI for BIM: each modality is processed into a vector space by its available technologies, and all are aligned into a shared space that serves the following learning. The main challenge comes from the limited research on encoding and aligning BIM-specific modalities, such as geometry and graphs, as current MLLMs primarily focus on vision, language, and audio (Lin et al., Reference Lin, Lee, Shoeybi, Lin, Catanzaro and Ping2024).

4.4.2. Learning patterns on graphs

Assuming efficient embedding of multimodal design data into graphs, learning-based algorithms can be applied to identify underlying patterns. Numerous existing Graph Neural Network (GNN) algorithms can be leveraged directly. GNNs can be viewed as a generalization of neural networks designed specifically to process unstructured graph data. The core operation of GNNs involves message passing, a method in which nodes iteratively exchange information with neighboring nodes to refine their representations that effectively capture the graph topology, such as GraphSAGE (Hamilton et al., Reference Hamilton, Ying and Leskovec2018). Moreover, Graph Convolutional Networks extend the concept of convolution to graph data, learning features by examining the attributes of neighboring nodes (Kipf and Welling, Reference Kipf and Welling2017). Graph Attention Networks introduce an attention mechanism, allowing the network to weigh the importance of different neighbors when aggregating information, thus focusing on the most relevant parts of the graph for each node (Veličković et al., Reference Veličković, Cucurull, Casanova, Romero, Liò and Bengio2018).

The effectiveness of GNNs has been demonstrated empirically across diverse domains. In biology, GNNs form a core component of AlphaFold, an AI program that has revolutionized protein folding prediction with high accuracy (Jumper et al., Reference Jumper, Evans, Pritzel, Green, Figurnov, Ronneberger, Tunyasuvunakool, Bates, Žídek, Potapenko and Hassabis2021). Compiling social networks as graphs enables the use of GNNs for various tasks, including community detection, identifying groups of users with strong interconnections, and link prediction, suggesting potential new connections between users (Ying et al., Reference Ying, He, Chen, Eksombatchai, Hamilton and Leskovec2018). In cybersecurity, they are used for detecting anomalies in computer networks by analyzing network graphs and identifying suspicious nodes or connections that might indicate malicious activity (King and Huang, Reference King and Huang2023).

The advantages of using graphs to represent design data directly align with the strengths of GNNs. Successful applications have shown the potential of GNNs to learn complex relationships in design models and successfully perform domain-specific tasks (Wang et al., Reference Wang, Sacks and Yeung2022; Xia et al., Reference Xia, Liao, Han, Zhang and Lu2025). While existing GNN architectures provide a robust foundation for application in BIGs, we recognize the unique characteristics of building design data, notably the variety of relationship types that are less common in domains such as protein interactions or social networks. Additionally, the differing contents and functions of subgraphs and the meta-graph require specialized attention. Consequently, there is a need to develop tailored learning algorithms within BIGs that explicitly address these distinctive data characteristics and optimize performance.

4.5. Training and dataset

Training methods in CV and NLP have evolved significantly, providing valuable insights. Three primary approaches are widely adopted: supervised learning, self-supervised learning, and transfer learning. Supervised learning relies on labeled datasets, such as ImageNet for vision tasks, to map input data to desired outputs. Self-supervised learning minimizes the need for labeled data by employing pretext tasks, like masked language modeling for NLP (Taylor, Reference Taylor1953). Transfer learning often involves using pretrained models that have been trained on large datasets and then are fine-tuned for specific tasks with smaller datasets. For example, YOLO models trained on the large COCO dataset have been fine-tuned using domain-specific datasets containing personal protective equipment to improve the accuracy of object classification (Nath et al., Reference Nath, Behzadan and Paal2020; Wang et al., Reference Wang, Wu, Yang, Thirunavukarasu, Evison and Zhao2021). These foundational training methods can be adapted to address diverse challenges in BIGs.

In BIGs, the training methods may initially focus on specific tasks due to the lack of large-scale datasets. For instance, algorithms could be trained for consistency maintenance between architectural and structural models or energy simulations using limited, task-specific data. However, as more comprehensive datasets become available, there could be a two-stage training paradigm similar to NLP and CV. In the first stage, a large and generic graph-based model could be pretrained on diverse BIM data, learning fundamental design knowledge and principles. This stage aims to equip the model with broad, generalizable knowledge, such as design intent and constraints, building codes and regulations, and so on. In the second stage, this pretrained model could be fine-tuned for specific applications, such as proposing suggestions for resolving conflicts, generative design, and so on. This paradigm would enable the efficient reuse of pretrained knowledge across tasks, accelerating the development of domain-specific solutions.

For example, generative design could become possible by fine-tuning the large graph-based models. In a manner similar to the use of RLHF in NLP, the fine-tuning could align outputs with designer preferences and domain requirements. A generative design assistant could help architects by predicting and completing partial designs. Designers might start by modeling a few walls and a floor, and the system could propose the rest of the layout based on learned patterns and feedback from the user. By incorporating human corrections and preferences iteratively, the model would refine its predictions, improving its ability to serve as an intelligent design assistant. This feedback loop would not only enhance generative tasks but also ensure alignment with real-world design practices and objectives.

However, we cannot ignore the difficulties of acquiring and constructing the dataset. Unlike the computer science domain, where sharing open data is common, the construction industry is more conservative due to concerns over intellectual property, competition, and data privacy. Design models often contain sensitive information, such as building layouts or client-specific details, raising security concerns if shared publicly.

Could using an intermediate representation like graphs ease part of the concerns? For example, by compiling design models into graphs, we can filter out all the sensitive information but preserve only essential design information. This abstraction reduces the risk of exposing proprietary or private data while retaining the core elements needed for research and development. If such a method were widely adopted, it might encourage more stakeholders in the construction industry to share anonymized or generalized models, fostering collaboration and enabling the creation of the large-scale datasets critical for advancing AI in building design.

4.6. Goals of learning

Graph learning tasks in building design will be tailored to support domain-specific applications by leveraging graph structures. Common tasks include node classification, link prediction, link classification, property prediction, clustering detection, graph classification, graph regression, and graph generation (Hamilton, Reference Hamilton2020; Wang et al., Reference Wang, Sacks, Ouyang, Ying and Borrmann2024). In the context of BIGs, the focus of graph learning is not limited to algorithmic development but also to addressing the practical needs of the construction industry.

For instance, link prediction is expected to automatically instantiate relationships among objects across design disciplines, forming linked graphs. Given that design models may contain thousands of objects, manually generating object links is impractical. Automating this process through graph learning is essential for achieving object-level multidisciplinary collaboration.

Similarly, object classification can predict the types of BIM objects, supporting interoperability across different software. Graph regression facilitates simulations on graphs, such as predicting energy performance. Graph generation enables the creation of new nodes and edges, contributing to generative design processes. These tasks exemplify how graph learning can address challenges that are difficult to solve with existing technologies. In short, the purpose of graph learning in BIGs is to serve discipline-specific design applications that can consider interdisciplinary dependencies.

4.7. Open questions

Considering the proposed graph learning framework in Figure 3, we identify four key stages: representation, learning (embedding and pattern recognition), training, and task applications. Consequently, we summarize the key research questions related to these stages as follows.

  1. Q6 What embedding methods can efficiently process multimodal BIM data (e.g., geometry, attributes, and relationships) and integrate them into a unified vector space while preserving semantic, contextual, and geometric information?

  2. Q7 What novel graph learning algorithms can be designed to fully leverage multimodal embeddings for the following analysis and prediction tasks, such as interdisciplinary link prediction, dynamic anomaly detection, and graph generation?

    Q8 How can we construct large BIG datasets?

  3. Q9 What training strategies can be devised to handle the computational demands of learning while serving domain-specific application scenarios?

5. Graph applications

In this section, we explore applications that are challenging to achieve with current file-based BIM solutions but become feasible through graph-based approaches. We categorize these applications into two groups. The first group focuses on leveraging explicit relationships in graphs to enhance collaboration and management functionalities, such as change propagation and graph-based version control (Table 1). Remodeling building information as graphs offers a promising solution to issues inherent in file-based BIM technologies. For these applications, generating explicit links is the main task. While link prediction has been validated using rule-based algorithms in pilot case studies (Wang et al., Reference Wang, Ouyang and Sacks2023), these algorithms face limitations in generalization and accuracy. To address these challenges, future research should explore learning-based algorithms, such as GNNs, which have the potential to improve link prediction performance by learning complex patterns.

Table 1. Graph applications to support collaboration, coordination, and management

* This application is not a fundamental module but a combination of change coordination and propagation, consistency maintenance, and graph-based version control.

The second group aims to unleash the possibilities of graph learning for generative design (Table 2). These applications heavily rely on advanced learning techniques, which are difficult to achieve using rule-based methods. Developing generative learning methods tailored to BIM is essential for implementing these applications. Success in this area could revolutionize modern design practices, enabling more intelligent and generative design processes.

Table 2. Applications of graph generative design

Notably, most applications are basic modules, except for concurrent design and engineering. Concurrent design integrates multiple modules, including change propagation, consistency maintenance, and version control. Research has demonstrated that transitioning workflows from sequential to concurrent can overlap design processes, shorten design cycles, and improve overall efficiency (Chen et al., Reference Chen, Lau, Yeung, Nyberg and Sacks2025). This makes concurrent design a critical area of focus for future development.

5.1. Change coordination and propagation

Some BIM software supports single-domain collaboration on the same model. For instance, Graphisoft (2024) supports multiple architects to work simultaneously on the same architectural model from different devices. Beyond a single discipline, cloud-based CDE platforms, such as Autodesk Cloud Collaborate (Autodesk, 2024) or Trimble Connect (Trimble, 2024), are among the most advanced solutions for multidisciplinary collaboration. These platforms rely on file-based workflows to manage design models, which cannot track and propagate design changes across disciplines. For instance, even just modifying the dimension of an architectural wall, the architect needs to export the whole project as files and mark the changes with screenshots or other methods. The file-based workflow requires extensive manual intervention and coordination.

To address these limitations, researchers have explored a graph-based approach by representing each discipline’s model as a subgraph and linking these subgraphs through instantiated multidisciplinary relationships (Törmä, Reference Törmä2013; Sacks et al., Reference Sacks, Wang, Ouyang, Utkucu and Chen2022). By leveraging the relationships, objects from different disciplines are no longer isolated. From the data perspective, the generated links allow data to be tracked and propagated from one subgraph to another. In a pilot case study, an architectural model created in Revit and a corresponding structural model designed in Tekla were compiled as separate architectural and structural subgraphs and hosted in the cloud. A geometry computation algorithm (Ouyang et al., Reference Ouyang, Wang and Sacks2023) was used to instantiate corresponding to and host/hostedby relationships, linking objects from the two subgraphs. For instance, hosting relationships were established between an architectural wall and two structural columns in Figure 4(a).

Figure 4. Change propagation from the architecture Revit to the Tekla structure by leveraging the interdisciplinary links (adopted and modified from Wang et al. (Reference Wang, Ying and Nyberg2024)).

When architects modified the design, such as moving an architectural wall westward (Figure 4(b)), the changed objects were computed and uploaded to the cloud server. Leveraging the established relationships, the system identified that the change would affect the structural objects hosted by the architectural wall and propagated the geometry of the wall to the structural software. As a result, engineers received the updated geometry reference directly from the cloud server, which was seamlessly loaded into their software without relying on file exchanges, as shown in Figure 4(c). This pilot study demonstrates the use of interdisciplinary relationships to link objects across design disciplines, enabling the propagation of relevant design changes on graphs.

5.2. Consistency maintenance

Researchers have explored interdisciplinary constraint relationships on linked graphs to maintain design consistency (Wang et al., Reference Wang, Ouyang and Sacks2023). This functionality extends beyond basic change coordination and propagation, as it requires computers to interpret design constraints and use these established relationships to resolve conflicts autonomously. To achieve this, Wang et al. (Reference Wang, Ouyang and Sacks2023) defined a set of constraint classes by abstracting expert knowledge. These were subsequently translated into graph formats. Figure 5 illustrates examples of design constraint classes between architectural slabs and structural slabs. For example, the first rule is that structural slabs must be fully contained within architectural slabs in the thickness dimension. For computational efficiency, bounding boxes are employed to evaluate constraints along the z-dimension for this rule. Each constraint instance has a binary “isFulfilled” property to indicate its status.

Figure 5. Examples of design constraint classes among architectural and structural slabs (adopted from Wang et al. (Reference Wang, Ouyang and Sacks2023)).

In a laboratory case study, the authors instantiated constraints using object geometry features and designed a mechanism to detect inconsistencies and propose solutions. Specifically, architects modified the model by moving walls and extending two architectural slabs. The cloud server identified the inconsistencies between the modified architectural slabs and the existing structural slabs. It then computed and suggested an automatic resolution by enlarging the structural slabs to comply with the predefined constraints, thereby resolving the conflict.

This laboratory study validates the concept of maintaining consistency on linked graphs using constraint relationships. However, it faces several limitations. First, a generic method for encapsulating design intent and defining constraint classes is desired. Second, the constraint instantiation algorithms were tailored to the specific case study, with a limited application scope. Third, the mechanism for resolving conflicts does not consider the consequences—if the solution will cause new inconsistencies.

5.3. Graph-based version control

Managing design changes and versions remains a significant challenge in BIM. Existing file-based systems require participants to package entire design models as files for sharing, even when only minor modifications occur. This approach leads to redundant data storage and inefficient communication, particularly in large and complex projects. Researchers have proposed graph-based approaches to detect and manage design changes at the object level, as shown in Figure 6.

Figure 6. Graph-based version control (adopted from Esser et al. (Reference Esser, Vilgertshofer and Borrmann2022)).

In the graph-based version control, design models are compared with previous versions on the sender side to detect changes. Consequently, only changes are transmitted and applied on the receiving side, enabling object-level synchronization (Esser et al., Reference Esser, Vilgertshofer and Borrmann2022). For example, two architects collaborate on the same architectural model. If one architect adds a window, only the modified object and corresponding relationships are detected and transmitted to the other architect’s model for synchronization, eliminating the need to transfer the entire file. Furthermore, the graph-based version control can support automated merging. Multiple users can work on the same project concurrently, each maintaining a development branch. Once their design work is complete, these branches can be merged into the master model after checking without conflicts (Esser et al., Reference Esser, Vilgertshofer and Borrmann2023). These examples illustrate the feasibility of graph-based version control for three key functionalities: (1) incremental change detection, (2) merging design branches, and (3) design variants and history tracking.

However, several challenges remain unresolved. First, there is a need for a standardized ontology to manage design changes and versions in a uniform and consistent way. It should contain version numbers, design timestamps, status (e.g., active or inactive), and so on. Second, resolving conflicts when merging different branches and variants requires further research. At present, when design conflicts occur, systems can display incremental changes, allowing users to review and decide whether to accept or reject specific design variants at the object level (Esser et al., Reference Esser, Vilgertshofer and Borrmann2023). A more advanced functionality would involve systems identifying conflicts and intelligently proposing solutions. This capability would enhance automation as the discussed consistency maintenance in Section 5.2.

5.4. Concurrent design and engineering

In building design, the prevailing design paradigm is sequential engineering, where design and development are executed step-by-step within single domains (Prasad, Reference Prasad1995). This often results in poor coordination, design inconsistencies, and inefficiencies in multidisciplinary collaboration (Akponeware and Adamu, Reference Akponeware and Adamu2017). Researchers in the AEC domain have advocated for adopting concurrent design and engineering, which shifts the sequential workflow into simultaneous collaboration among multidisciplinary teams (Anumba et al., Reference Anumba, Baron and Duke1997; Anumba et al., Reference Anumba, Baugh and Khalfan2002). However, the inherent fragmentation of the construction industry poses significant challenges to implementing concurrent design.

Graph-based applications for change coordination and version control have demonstrated the feasibility of two essential features for enabling concurrent design: (1) maintaining domain-specific models while receiving synchronized updates of other disciplines from the cloud; (2) coordinating relevant changes across disciplines and software platforms; and (3) managing design changes at the object level. In addition, the concurrent design solution should furthermore be equipped with a communication platform where participants can be involved to communicate and comment on design models with precise locations. Additionally, this technology should support real-time collaboration and coordination, allowing participants to review the latest models and changes from any discipline and platform.

In practice, high transparency and synchronized collaboration may raise concerns among users, as partially developed designs are continuously accessible to others. To address this, a permission mechanism should enable users to control the visibility of specific model statuses, allowing them to manage what others can view. While this may limit the full real-time collaboration, it can still overlap workflows to shorten design cycles and improve overall efficiency compared to sequential design (Chen et al., Reference Chen, Lau, Yeung, Nyberg and Sacks2025).

5.5. Design issue detection

As design models are represented as graphs, another potential application is detecting design issues through anomaly detection. This can be achieved by identifying anomalous patterns, nodes, edges, or subgraphs on graphs. Graph anomaly detection has already shown significant success in other domains. For instance, in finance, subgraph-level anomaly detection is used on transaction graphs to identify patterns associated with fraudulent activities, such as money laundering (Pei et al., Reference Pei, Lyu, van Ipenburg and Pechenizkiy2021). In biology, detecting node-level anomalies in protein–protein interaction networks helps reveal potential causal genes for rare diseases (Kuzmanov and Emili, Reference Kuzmanov and Emili2013). Similarly, in social networks, user interactions are compiled as graphs to detect unusual behaviors, such as spam activity and internet attacks (Lamichhane and Eberle, Reference Lamichhane and Eberle2024).

Applying anomaly detection to BIGs can enable the system to monitor dynamic graphs and identify design issues, such as clashes, code violations, improper object placements, or unusual layouts deviating from established design patterns. Unlike existing studies that classify partial model graphs for code compliance (Bloch et al., Reference Bloch, Borrmann and Pauwels2023), this approach would operate on the entire design graph and dynamically detect anomalies.

In the future, anomaly detection could focus on identifying more abstract anomalies coupled with instantiated constraint relationships to ensure issue-free and code-compliant designs. Constraints provide explicit relationships for the system to validate designs, but not all design intents and constraints can be instantiated by rules. Anomaly detection can supplement this by identifying abstract patterns, such as unconventional spatial layouts, that are difficult to encapsulate as explicit languages. By combining constraints with advanced anomaly detection, BIGs could enable more comprehensive and intelligent design issue detection.

5.6. Graph-based life cycle data management

Information Containers for Linked Document Delivery (ICDD) aims to define an open and stable container format for exchanging heterogeneous files to facilitate the delivery, storage, and archiving of documents that describe an asset throughout its entire life cycle (ISO, 2020). The LBD community leverages semantic web technologies to implement the ICDD concept to manage design files in the life cycle (Senthilvel et al., Reference Senthilvel, Oraskari and Beetz2020).

The ICDD utilizes containers, which are structured collections of documents and graphs that link these documents (Göbels and Beetz, Reference Göbels and Beetz2022; Hagedorn et al., Reference Hagedorn, Senthilvel, Schevers and Verhelst2023). Documents could be arbitrary and heterogeneous, such as 2D CAD drawings, 3D BIM models, construction schedules, and external links, which are organized into folders with a hierarchy (Hagedorn et al., Reference Hagedorn, Senthilvel, Schevers and Verhelst2023).

Instead, the ICDD container can be conceptualized as a graph, where nodes represent files or objects, and edges represent relationships between them. Based on different granularities, links can be placed between files and objects (Hagedorn et al., Reference Hagedorn, Senthilvel, Schevers and Verhelst2023). For instance, linking 3D model files with 2D CAD drawings can enhance coordination. Researchers also proposed methods to link objects within files to objects in another file, such as linking the same wall in 2D and 3D models for an interactive evaluation (Borrmann et al., Reference Borrmann, Abualdenien and Krijnen2021).

The concept of ICDD is straightforward to leverage graphs to integrate and manage data via a project’s life cycle. However, the real challenges come from the methods to instantiate relationships. Existing studies illustrated the way of generating links by using the same identifier (Borrmann et al., Reference Borrmann, Abualdenien and Krijnen2021; Hagedorn et al., Reference Hagedorn, Pauwels and König2023). Beyond that, many links are still generated manually (Fuchs and Scherer, Reference Fuchs and Scherer2017).

5.7. Design completion

Design completion aims to assist architects and engineers in completing their designs. It could be similar to existing AI code assistant tools, such as GitHub Copilot (GitHub, 2025), which can complete code lines or generate code from descriptions. Technically, design completion can be regarded as a task of predicting nodes and edges based on existing graphs. For instance, when an architect begins designing an apartment by placing a floor and two walls (an initial graph), the system can predict the remaining elements of the room, such as walls, windows, and doors, by generating nodes and links. It can extend to varying graph depths. At one-step depth, the system predicts directly connected nodes, such as walls that complete the room boundary or windows that fit into existing walls. At multistep depth, the system predicts more complex objects and relationships, such as objects in adjacent rooms.

5.8. Detailing generation

Detailing generation transforms low-LOD designs into high-LOD representations by leveraging graph learning. This process can be framed as two core tasks: graph matching and graph substitution. Graph matching focuses on identifying and classifying low-LOD nodes or subgraphs within the graph. Graph substitution involves replacing low-LOD components with detailed high-LOD counterparts, which can be achieved in two different approaches. The first one is a similarity-based substitution. Using a preconstructed library of high-LOD subgraphs, substitution becomes a task of similarity computation. Alternatively, high-LOD designs can be generated directly using generative models.

Existing has explored to achieve the detailing design by transplanting high LOD BIM objects to lower ones (Jang and Lee, Reference Jang and Lee2024). It requires users to input both a low LOD model as well as a high LOD project model. The matching algorithm mainly utilizes object properties and class types from low LOD models to predict the potential object classes in high LOD models for substitution (Jang and Lee, Reference Jang and Lee2024). The framework heavily relies on the quality and relevance of the high LOD model. Misaligned input models can lead to mismatched substitutions (Jang and Lee, Reference Jang and Lee2024). Additionally, the matching algorithm focuses on properties without considering the broader relational context of objects (e.g., functional dependencies or connections), resulting in inaccuracies when dealing with complex architectural contexts, such as layered walls requiring waterproofing (Jang and Lee, Reference Jang and Lee2024).

On the other hand, graphs show the potential of addressing these limitations. Graphs naturally encode relationships that can provide the relational context for algorithms to learn. Graphs are scalable to represent and store different LOD objects and subgraphs, even for large and complex BIM models. Additionally, the dynamic property of graphs can iteratively update graphs and refine substitutions over time.

5.9. Paired design

Paired design generation focuses on generating models for one discipline based on another. For example, a structural design could be generated from an architectural model. BIGs, with the ability to encode linked graphs, have the potential to extrapolate the entire structural graph from the architectural graph. The system could recommend multiple structural design options, ensuring that generated models are conflict-free and compliant with codes. Furthermore, this capability could extend to generating MEP designs after completing architectural and structural models.

Paired design generation can be formed as a conditional graph generation task, using a full graph as an input to generate another full graph. It could be a challenging graph learning task, as a similar scenario is even rarer in the computer science domain. In theory, the graphs in BIGs are highly interrelated, providing the basis for algorithms to learn. One possible direction is to develop Variational Graph Auto-encoders that can learn a latent representation of architectural graphs and decode it into structural graphs.

In a recent study, Xia et al. (Reference Xia, Liao, Han, Zhang and Lu2025) illustrated a simple approach to paired design generation by formulating the problem as an edge classification task. In their approach, they generate structural shear wall and beam layouts from architectural designs. The input is an architectural graph generated from the architectural CAD drawings by modeling walls, doors, and windows as edges in the graph (Xia et al., Reference Xia, Liao, Han, Zhang and Lu2025). The basic assumption is that structural elements must be embedded in corresponding architectural elements (e.g., structural columns must be located within architectural walls). Thus, the authors modeled the structural layout design as a link classification task on architectural graphs—predicting whether the architectural elements (edges) should be structural elements. They applied a GraphSAGE model for link classification, followed by a rule-based method for refining the GNN-generated layouts. The approach is too simple for application in practice, because the well-developed architectural graphs used as input are not representative of the starting point for conceptual design that structural engineers must use. In addition, the one-to-one relationship between architectural and structural elements does not hold in the general case. Nevertheless, the work suggests that a paired design may be feasible in the future.

5.10. Generative design from text

Another one is to generate design models from user descriptions directly. The input only requires requirements and description of this project, such as the type of this project, size and scale, usage, site information, client needs, and so on. It should analyze and understand the requirements and generate corresponding design graphs.

Existing studies have achieved the text-to-BIM generative design by leveraging LLMs, where LLMs are used to understand the requirements and generate programming codes that can call the BIM software Application Programming Interfaces (APIs) for constructing 3D models (Du et al., Reference Du, Esser, Nousias and Borrmann2024). This further expands the applications of LLMs in the BIM domain, but it meets the challenges of only generating regular, noncurved building models in the early design stage (Du et al., Reference Du, Esser, Nousias and Borrmann2024). It also requires human validation to ensure compliance with design requirements (Du et al., Reference Du, Esser, Nousias and Borrmann2024). These limitations stem from the fact that LLMs are designed for processing natural language in a tokenized format, with limited capacity to understand the complexities of building design. By contrast, graph-based systems have the potential to learn and represent design more effectively, offering possibilities for advanced generative design applications.

Zhao et al. (Reference Zhao, Fei, Huang, Feng, Liao and Lu2023) explored the use of graphs and textual descriptions to generate structural plans. The authors represented architectural CAD drawings as graphs and considered three design conditions: (1) building height, (2) peak ground acceleration of the design basis earthquake, and (3) characteristic ground period. These conditions were encoded as numerical vectors, concatenated with the original node and edge features, and then input into a GNN model for learning (Zhao et al., Reference Zhao, Fei, Huang, Feng, Liao and Lu2023). This study highlights the potential of integrating building elements with relationships and design descriptions to enhance generative design capabilities. However, this approach relies on a predefined architectural graph as input and is restricted to shear wall layout generation. Furthermore, it processes only three specific parameters, limiting its ability to generalize to complex design scenarios with diverse constraints. Incorporating LLM-based methods could enable the processing and embedding of a wider range of textual design descriptions, thereby improving scalability and adaptability. Nonetheless, the integration of LLMs with other techniques for efficiently embedding multimodal design data (text, relationships, geometries, etc.) remains unexplored, posing a critical and fundamental research challenge in BIGs, as discussed in Section 4.4.

5.11. Open questions

This section explores applications that are challenging for traditional file-based BIM techniques but become feasible after adopting graphs. Some of them have been validated in recent studies, such as change coordination, consistency maintenance, and graph-based version control. However, behind the validation, some engineering and research questions still remain, such as graph efficiency and complexity. Furthermore, we discuss several applications that heavily leverage graph learning, highlighting the development of advanced graph learning techniques as crucial for realizing these intelligent applications. We summarize the open research questions as follows:

  • Q10 What is the performance of linked graphs with the latest technologies in real-world scenarios in terms of efficiency and complexity?

  • Q11 How can learning algorithms efficiently identify and resolve inconsistencies (e.g., geometric clashes or code violations) while incorporating explainable AI techniques for transparency?

  • Q12 What graph learning methods can be developed to support various levels of generative design, from design completion and detailing generation to generation from text?

  • Q13 How can graph anomaly detection techniques be tailored to detect unconventional patterns, design inconsistencies, or code violations, and how can they integrate established domain-specific constraint relationships?

  • Q14 What other practical applications can be developed by leveraging graphs and graph learning? For example, can BIGs support the automatic generation of bills of quantities and construction schedules? Might they be used to compile data for simulations of building performance or of production progress, both in the context of building digital twins (Sacks et al., Reference Sacks, Brilakis, Pikas, Xie and Girolami2020)?

6. Conclusion

Despite advances in BIM technologies and processes, the AEC industry continues to face significant challenges. Existing file-based workflows require teams to exchange design models in files across disciplines and software, often resulting in inconsistencies and cumbersome version management. The limitations of current data schemas hinder the accurate and comprehensive encapsulation of design and constraints, causing interoperability issues and difficulty for computers to understand the design. Additionally, current AI studies in BIM are constrained by sampling partial modeling data as required formats for the adopted mature algorithms, which fail to consider the richness and multimodality of building information.

We propose BIGs as an alternative method to model design information into a graph-based format and store it in a centralized database, to support graph-based applications and facilitate graph learning. We define a BIG as a collection of discipline-specific subgraphs linked by interdisciplinary relationships in a meta-graph. In a BIG, nodes represent objects, and edges represent relationships between objects within and across disciplines. The relationships are both intradisciplinary and interdisciplinary. BIGs are designed to support computation and domain applications with properties, such as attributes, multimodality, and computability. As such, BIGs integrate siloed design data, bridging the information generated by different professional disciplines that collaborate to design buildings.

Graphs are eminently suited to storing relationships between objects, allowing for more granular object-level data representation. This introduces a new dimension of contextual data that can capture design intent and constraints among building objects. By leveraging explicit representation of object relationships, various innovative applications that are hard to achieve using current BIM technologies–such as multidisciplinary change propagation, automated consistency maintenance, graph-level version control, and more–become feasible and have been validated in recent studies.

Furthermore, BIGs are designed for computation and learning. BIGs preserve multimodal design data, including geometry, attributes, and topology, which can be embedded into high-dimensional vector spaces in a learning-suitable format. Following that, advanced learning algorithms can leverage these embeddings to identify statistical patterns and support downstream tasks, such as link prediction, anomaly detection, and graph generation. These graph-based tasks, in turn, serve for domain-specific applications, including generative design.

In this article, we have identified and discussed three core challenges of BIGs: (1) Encapsulating object relationships, especially for design intent and constraints. (2) Designing a series of graph learning techniques, from embedding methods, to learning algorithms, to training strategies, and dataset construction. (3) Developing applications on graphs that can leverage objects and relationships to enable collaborative and intelligent design and construction functionalities. As an opinion article, this work does not provide mature solutions, but rather outlines critical research directions with the intent to inspire further investigation. To facilitate ongoing research and industry engagement, we have summarized the key open questions for each of these three challenges.

The proposition of BIGs makes three specific contributions to research and to industry. First, we define a method to remodel building information by structuring multidisciplinary design information in linked graphs, capturing object relationships. BIGs provide a more interconnected and granular approach than information exchange with files, enabling seamless data integration across disciplines. Second, we introduce the use of graphs as the data representation that not only preserves the rich, complex, and multimodal nature of building information but also facilitates machine learning. This can bridge the gap between applying machine learning to BIM and unleash the potential of AI in the AEC domain. Third, we outline various graph-based applications, demonstrating how BIGs can enable software vendors to overcome the limitations of existing technologies and unlock new possibilities in BIM tools for design, collaboration, life cycle management, and more.

We believe that BIGs mark the beginning of a paradigm shift in how building design data is leveraged for domain applications and computational learning. We acknowledge that this work serves only as an initial exploration, and we encourage researchers and industry partners to join the journey. Through intensive conceptual development of BIGS and devising new machine learning applications for them, researchers in academia and industry can deepen our understanding of BIGs and unlock their potential for a more collaborative, intelligent, and generative design future.

Data availability statement

No data were used or generated in this research.

Acknowledgments

The authors thank Professor André Borrmann for his thorough internal review and insightful, valuable feedback.

Author contribution

Conceptualization: Z.W. and R.S. Methodology: Z.W. and R.S. Data visualization: Z.W. Writing original draft: Z.W. and R.S. All authors approved the final submitted draft.

Competing interests

The authors declare none.

Ethical standard

The research meets all ethical guidelines, including adherence to the legal requirements of the study country.

Footnotes

1 Professional practice and commercial liability dictate that changes to an object in a BIM model should only be made by either the designer who created them or with that designer’s explicit permission.

References

Afsari, K, Eastman, CM and Castro-Lacouture, D (2017) JavaScript object notation (JSON) data serialization for IFC schema in web-based BIM data exchange. Automation in Construction 77, 2451. https://doi.org/10.1016/j.autcon.2017.01.011.Google Scholar
Akponeware, AO and Adamu, ZA (2017) Clash detection or clash avoidance? An investigation into coordination problems in 3D BIM. Buildings 7(3). https://doi.org/10.3390/buildings7030075.CrossRefGoogle Scholar
Anumba, C, Baron, G and Duke, A (1997) Information and communications technologies to facilitate concurrent engineering in construction. BT Technology Journal 15(3), 199207.CrossRefGoogle Scholar
Anumba, CJ, Baugh, C and Khalfan, MM (2002) Organisational structures to support concurrent engineering in construction. Industrial Management & Data Systems 102(5), 260270.CrossRefGoogle Scholar
Augenbroe, G (1994a) Integrated building Design Systems in Context of product data technology. Journal of Computing in Civil Engineering 8(4), 420435. https://doi.org/10.1061/(ASCE)0887-3801(1994)8:4(420.CrossRefGoogle Scholar
Augenbroe, G (1994b) Integrated use of building design tools: Results from the combine project. In Automation Based Creative Design–Research and Perspectives (pp. 205218). Elsevier.CrossRefGoogle Scholar
Autodesk. (2024). Autodesk Cloud Collaborate. Retrieved January 14, 2024, from https://www.autodesk.ae/products/bim-collaborate/overview?term=1-YEAR&tab=subscriptionGoogle Scholar
Beetz, J (2009) Facilitating Distributed Collaboration in the AEC/FM Sector Using Semantic Web Technologies (Publication No. 2009) [Doctoral dissertation. Eindhoven: Technische Universiteit Eindhoven. https://doi.org/10.6100/IR652808.Google Scholar
Belsky, M, Sacks, R and Brilakis, I (2016) Semantic enrichment for building information modeling. Computer-Aided Civil and Infrastructure Engineering 31(4), 261274. https://doi.org/10.1111/mice.12128.CrossRefGoogle Scholar
Björk, B-C (1989) Basic structure of a proposed building product model. Computer-Aided Design 21(2), 7178. https://doi.org/10.1016/0010-4485(89)90141-3.CrossRefGoogle Scholar
Björk, B-C (1994) RATAS project—Developing an infrastructure for computer-integrated construction. Journal of Computing in Civil Engineering 8(4), 401419. https://doi.org/10.1061/(ASCE)0887-3801(1994)8:4(401.CrossRefGoogle Scholar
Bloch, T (2022) Connecting research on semantic enrichment of BIM-review of approaches, methods and possible applications. J. Inf. Technol. Constr. 27, 416440.Google Scholar
Bloch, T, Borrmann, A and Pauwels, P (2023) Graph-based learning for automated code checking– Exploring the application of graph neural networks for design review. Advanced Engineering Informatics 58, 102137. https://doi.org/10.1016/j.aei.2023.102137.Google Scholar
Bloch, T and Sacks, R (2018) Comparing machine learning and rule-based inferencing for semantic enrichment of BIM models. Automation in Construction 91, 256272. https://doi.org/10.1016/j.autcon.2018.03.018CrossRefGoogle Scholar
Bollobás, B (2013) Modern Graph Theory, Vol. 184. Springer Science & Business Media.Google Scholar
Bonduel, M, Oraskari, J, Pauwels, P, Vergauwen, M and Klein, R (2018) The IFC to linked building data converter: Current status. In 6th International Workshop on Linked Data in Architecture and Construction, pp. 3443.Google Scholar
Bonduel, M, Wagner, A and Pauwels, P (2020) FOG: File Ontology for Geometry Formats. https://mathib.github.io/fog-ontology/#propertiesGoogle Scholar
Bondy, JA and Murty, USR (2008) Graph Theory. Springer Publishing Company, Incorporated.CrossRefGoogle Scholar
Borrmann, A, Abualdenien, J and Krijnen, T (2021) Information containers providing deep linkage of drawings and BIM models. In Proceedings of the CIB W78 Conference 2021.Google Scholar
Borrmann, A, Hyvärinen, J and Rank, E (2009) Spatial constraints in collaborative design processes. In Proceedings of International Conference on Intelligent of Computing in Engineering (ICE’09).Google Scholar
Borrmann, A, König, M, Koch, C and Beetz, J (2018, September) Building Information Modeling (Borrmann, A, König, M, Koch, C, and Beetz, J, Eds.). Springer International Publishing. https://doi.org/10.1007/978-3-319-92862-3CrossRefGoogle Scholar
buildingSMART (2024, December) The evolution of IFC: The Path to IFC5. https://www.buildingsmart.es/2024/12/03/the-evolution-of-ifc-the-path-to-ifc5/Google Scholar
buildingSMART (2025c) Industry Foundation Classes (IFC). https://technical.buildingsmart.org/standards/ifc/.Google Scholar
Chen, S, Lau, A, Yeung, T, Nyberg, K and Sacks, R (2025) Comparing concurrent and sequential engineering in cloud-BIM interdisciplinary design collaboration: An experimental study (accepted for publication). Architectural Engineering and Design Management. https://doi.org/10.1080/17452007.2025.2549493CrossRefGoogle Scholar
Collins, FC, Braun, A, Ringsquandl, M, Hall, DM and Borrmann, A (2021) Assessing IFC classes with means of geometric deep learning on different graph encodings. Proceedings of the 2021 European Conference on Computing in Construction 2, 332341. https://doi.org/10.35490/ec3.2021.168.CrossRefGoogle Scholar
Crawl, C (2024, November). A Free, Open Repository of Web Crawl Data that Can Be Used by Anyone.Google Scholar
Daum, S and Borrmann, A (2014) Processing of topological BIM queries using boundary representation based methods. Advanced Engineering Informatics 28(4), 272286. https://doi.org/10.1016/j.aei.2014.06.001.CrossRefGoogle Scholar
de Wilde, P (2019) Ten questions concerning building performance analysis. Building and Environment 153, 110117. https://doi.org/10.1016/j.buildenv.2019.02.019.CrossRefGoogle Scholar
DeepSeek-AI, Guo, D, Yang, D, Zhang, H, Song, J, Zhang, R, Xu, R, Zhu, Q, Ma, S, Wang, P, Bi, X, Zhang, X, Yu, X, Wu, Y, Wu, ZF, Gou, Z, Shao, Z, Li, Z, Gao, Z, … Zhang, R (2025) DeepSeek-R1: Incentivizing Reasoning Capability in LLMs Via Reinforcement Learning. https://arxiv.org/abs/2501.12948Google Scholar
Deng, J, Dong, W, Socher, R, Li, L-J, Li, K and Fei-Fei, L (2009) Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp. 248255.CrossRefGoogle Scholar
Du, C, Esser, S, Nousias, S and Borrmann, A (2024) Text2BIM: Generating Building Models Using a Large Language Model-Based Multi-Agent Framework. https://arxiv.org/abs/2408.08054Google Scholar
Eastman, C (1975) The use of computers instead of drawings in building design. AIA Journal 63(3), 4650.Google Scholar
Esser, S, Vilgertshofer, S and Borrmann, A (2022) Graph-based version control for asynchronous BIM collaboration. Advanced Engineering Informatics 53, 101664. https://doi.org/10.1016/j.aei.2022.101664.Google Scholar
Esser, S, Vilgertshofer, S and Borrmann, A (2023) Version control for asynchronous BIM collaboration: Model merging through graph analysis and transformation. Automation in Construction 155, 105063. https://doi.org/10.1016/j.autcon.2023.105063.CrossRefGoogle Scholar
Fuchs, S and Scherer, RJ (2017) Multimodels — Instant nD-modeling using original data. Automation in Construction 75, 2232. https://doi.org/10.1016/j.autcon.2016.11.013.CrossRefGoogle Scholar
GitHub. (2025). GitHub Copilot · your AI Pair Programmer. https://github.com/features/copilotGoogle Scholar
Göbels, A and Beetz, J (2022) Using ICDD containers for documentation data archives: Semi-automated creation of linked documentation data archives using an information container for linked document delivery. Proceedings of the 27th International Conference on Cultural Heritage and New Technologies.Google Scholar
Graphisoft. (2024). Archicad. Retrieved December 14, 2024, from https://graphisoft.com/uk/Google Scholar
Hagedorn, P, Pauwels, P and König, M (2023) Semantic rule checking of cross-domain building data in information containers for linked document delivery using the shapes constraint language. Automation in Construction 156, 105106. https://doi.org/10.1016/j.autcon.2023.105106.CrossRefGoogle Scholar
Hagedorn, P, Senthilvel, M, Schevers, H and Verhelst, L (2023) Towards Usable ICDD Containers for Ontology-Driven Data Linking and Link Validation. LDAC, pp. 3546.Google Scholar
Hamilton, WL (2020) Graph Representation Learning. Morgan & Claypool Publishers.CrossRefGoogle Scholar
Hamilton, WL, Ying, R and Leskovec, J (2018) Inductive Representation Learning on Large Graphs. https://arxiv.org/abs/1706.02216Google Scholar
He, K, Zhang, X, Ren, S and Sun, J (2016) Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770778.Google Scholar
Hinton, G, Vinyals, O and Dean, J (2015) Distilling the knowledge in a neural network. Preprint, arXiv:1503.02531.Google Scholar
Ingram, J (2020) Understanding BIM: The Past, Present and Future. Routledge.Google Scholar
Inhan Kim, JK and Seo, J (2012) Development of an IFC-based IDF converter for supporting energy performance assessment in the early design phase. Journal of Asian Architecture and Building Engineering 11(2), 313320. https://doi.org/10.3130/jaabe.11.313.Google Scholar
ISO. (2020). Information container for linked document delivery. In Iso 21597–1:2020. https://www.iso.org/standard/74389.htmlGoogle Scholar
ISO 16739-1. (2018). Industry Foundation Classes (IFC) for Data Sharing in the Construction and Facility Management Industries — Part 1: Data Schema [ISO]. https://www.iso.org/standard/70303.htmlGoogle Scholar
ISO19650-1. (2018). Organization and Digitization of Information about Buildings and Civil Engineering Works, Including Building Information Modelling (BIM) — Information Management Using Building Information Modelling — Part 1: Concepts and Principles (1st Ed.). ISO. https://www.iso.org/standard/68078.html#lifecycleGoogle Scholar
Jang, S and Lee, G (2024) BIM library transplant: Bridging human expertise and artificial intelligence for customized design detailing. Journal of Computing in Civil Engineering 38(2), 04024004. https://doi.org/10.1061/JCCEE5.CPENG-5680.CrossRefGoogle Scholar
Ji, Y, Borrmann, A, Beetz, J and Obergrießer, M (2013) Exchange of parametric bridge models using a neutral data format. Journal of Computing in Civil Engineering 27(6), 593606. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000286.CrossRefGoogle Scholar
Jumper, J, Evans, R, Pritzel, A, Green, T, Figurnov, M, Ronneberger, O, Tunyasuvunakool, K, Bates, R, Žídek, A, Potapenko, A, … Hassabis, D (2021) Highly accurate protein structure prediction with AlphaFold. Nature 596(7873), 583589.CrossRefGoogle ScholarPubMed
King, IJ and Huang, HH (2023) Euler: Detecting network lateral movement via scalable temporal link prediction. ACM Transactions on Privacy and Security 26(3), 136.CrossRefGoogle Scholar
Kipf, TN and Welling, M (2017) Semi-supervised classification with graph convolutional networks. https://arxiv.org/abs/1609.02907Google Scholar
Koo, B, Jung, R and Yu, Y (2021) Automatic classification ofwall and doorBIM element subtypes using 3D geometric deep neural networks. Advanced Engineering Informatics 47, 101200. https://doi.org/10.1016/j.aei.2020.101200CrossRefGoogle Scholar
Koo, B, La, S, Cho, NW and Yu, Y (2019) Using support vector machines to classify building elements for checking the semantic integrity of building information models. Automation in Construction 98, 183194. https://doi.org/10.1016/j.autcon.2018.11.015.CrossRefGoogle Scholar
Kuzmanov, U and Emili, A (2013) Protein-protein interaction networks: Probing disease mechanisms using model systems. Genome Medicine 5, 112.CrossRefGoogle ScholarPubMed
Lamichhane, PB and Eberle, W (2024) Anomaly Detection in Graph Structured Data:ASurvey. https://arxiv.org/abs/2405.06172Google Scholar
LBD. (2024, December). Linked Building Data Community. https://www.w3.org/community/lbd/Google Scholar
LeCun, Y, Bottou, L, Bengio, Y and Haffner, P (1998) Gradient-based learning applied to document recognition. Proceedings of the IEEE 86(11), 22782324.CrossRefGoogle Scholar
Lin, S-C, Lee, C, Shoeybi, M, Lin, J, Catanzaro, B and Ping, W (2024) MM-Embed: Universal Multimodal Retrieval with Multimodal LLMs. https://arxiv.org/abs/2411.02571Google Scholar
Liu, Y, Wang, K, Shao, W, Luo, P, Qiao, Y, Shou, MZ, Zhang, K and You, Y (2024) MLLMs-Augmented Visual-Language Representation Learning. https://arxiv.org/abs/2311.18765Google Scholar
Mikolov, T (2013) Efficient Estimation of Word Representations in Vector Space arXiv preprint arXiv:1301.3781, 3781.Google Scholar
Nath, ND, Behzadan, AH and Paal, SG (2020) Deep learning for site safety: Real-time detection of personal protective equipment. Automation in Construction 112, 103085. https://doi.org/10.1016/j.autcon.2020.103085.CrossRefGoogle Scholar
OpenAI. (2022a, January). Aligning Language Models to Follow Instructions. https://openai.com/index/instruction-following/Google Scholar
OpenAI. (2022b, November). Introducing ChatGPT. https://openai.com/index/chatgpt/Google Scholar
OpenAI. (2024). Learning to Reason with LLMs. https://openai.com/index/learning-to-reason-with-llms/Google Scholar
Ouyang, B (2023) Foundations of Collaborative Multi-Discipline Cloud Building Information Modeling [Master’s thesis, Technion Israel Institute of Technology].Google Scholar
Ouyang, B, Wang, Z and Sacks, R (2023) Semantic enrichment of object associations across federated BIM semantic graphs in a common data environment. In Ecppm 2022-Ework and Ebusiness in Architecture, Engineering and Construction 2022 (pp. 591598). CRC Press.CrossRefGoogle Scholar
Pauwels, P (2014) Supporting decision-making in the building life-cycle using linked building data. Buildings 4(3), 549579.CrossRefGoogle Scholar
Pauwels, P, Costin, A and Rasmussen, MH (2022) Knowledge Graphs and Linked Data for the Built Environment. Cham: Springer. https://doi.org/10.1007/978-3-030-82430-3_7.CrossRefGoogle Scholar
Pauwels, P and Terkaj, W (2016) EXPRESS to OWL for construction industry: Towards a recommendable and usable ifcOWL ontology. Automation in Construction 63, 100133. https://doi.org/10.1016/j.autcon.2015.12.003.CrossRefGoogle Scholar
Pei, Y, Lyu, F, van Ipenburg, W and Pechenizkiy, M (2021) Subgraph anomaly detection in financial transaction networks. Proceedings of the First ACM International Conference on AI in Finance. https://doi.org/10.1145/3383455.3422548.Google Scholar
Prasad, B (1995) Sequential versus concurrent engineering—An analogy. Concurrent Engineering 3(4), 250255.CrossRefGoogle Scholar
Radford, A, Kim, JW, Hallacy, C, Ramesh, A, Goh, G, Agarwal, S, Sastry, G, Askell, A, Mishkin, P, Clark, J, Krueger, G and Sutskever, I (2021) Learning transferable visual models from natural language supervision. https://arxiv.org/abs/2103.00020Google Scholar
Rasmussen, MH, Lefrançois, M, Schneider, GF and Pauwels, P (2020) BOT: The building topology ontology of the W3C linked building data group. In Janowicz, K (ed), Semantic Web, Vol. 12, pp. 143161. https://doi.org/10.3233/SW-200385.CrossRefGoogle Scholar
Russakovsky, O, Deng, J, Su, H, Krause, J, Satheesh, S, Ma, S, Huang, Z, Karpathy, A, Khosla, A, Bernstein, M, … Fei-Fei, L (2015) ImageNet large scale visual recognition challenge. International Journal of Computer Vision 115, 211252.CrossRefGoogle Scholar
Sacks, R, Brilakis, I, Pikas, E, Xie, HS and Girolami, M (2020) Construction with digital twin information systems. Data-Centric Engineering 1, e14.CrossRefGoogle Scholar
Sacks, R, Lee, G, Burdi, L and Bolpagni, M (2025). BIM Handbook: A Guide to Building Information Modeling for Owners, Designers, Engineers, Contractors and Facility Managers, 4th Edition, John Wiley and Sons, Hoboken NJ, p. 544. ISBN 978-1-394-22222-3.CrossRefGoogle Scholar
Sacks, R, Eastman, CM and Lee, G (2004) Parametric 3D modeling in building construction with examples from precast concrete. Automation in Construction 13(3), 291312. https://doi.org/10.1016/S0926-5805(03)00043-8.CrossRefGoogle Scholar
Sacks, R, Wang, Z, Ouyang, B, Utkucu, D and Chen, S (2022) Toward artificially intelligent cloud-based building information modelling for collaborative multidisciplinary design. Advanced Engineering Informatics 53, 101711. https://doi.org/10.1016/j.aei.2022.101711CrossRefGoogle Scholar
Senthilvel, M, Oraskari, J and Beetz, J (2020) Common Data Environments for the Information Container for Linked Document Delivery. LDAC, pp. 132145.Google Scholar
Taylor, WL (1953) Cloze procedure”: A new tool for measuring readability. Journalism Quarterly 30(4), 415433.Google Scholar
Törmä, S (2013) Semantic linking of building information models. In Proceedings - 2013 IEEE 7th International Conference on Semantic Computing, ISC 2013. IEEE, pp. 412419. https://doi.org/10.1109/ICSC.2013.80Google Scholar
Trimble. (2024). Trimble connect. Retrieved January 14, 2024, from https://connect.trimble.com/Google Scholar
Vaswani, A (2017) Attention is all you need. Advances in Neural Information Processing Systems.Google Scholar
Veličković, P, Cucurull, G, Casanova, A, Romero, A, Liò, P and Bengio, Y (2018) Graph attention networks. https://arxiv.org/abs/1710.10903.Google Scholar
Venugopal, M, Eastman, CM and Teizer, J (2015) An ontology-based analysis of the industry foundation class schema for building information model exchanges [collective intelligence modeling, analysis, and synthesis for innovative engineering decision making special issue of the 1st international conference on civil and building engineering informatics]. Advanced Engineering Informatics 29(4), 940957. https://doi.org/10.1016/j.aei.2015.09.006.CrossRefGoogle Scholar
Wang, Z, Ouyang, B and Sacks, R (2023) Graph-based inter-domain consistency maintenance for BIM models. Automation in Construction 154, 104979. https://doi.org/10.1016/j.autcon.2023.104979.CrossRefGoogle Scholar
Wang, Z, Sacks, R, Ouyang, B, Ying, H and Borrmann, A (2024) A framework for generic semantic enrichment of BIM models. Journal of Computing in Civil Engineering 38(1), 04023038.CrossRefGoogle Scholar
Wang, Z, Sacks, R and Yeung, T (2022) Exploring graph neural networks for semantic enrichment: Room type classification. Automation in Construction 134, 104039. https://doi.org/10.1016/j.autcon.2021.104039.CrossRefGoogle Scholar
Wang, Z, Wu, Y, Yang, L, Thirunavukarasu, A, Evison, C and Zhao, Y (2021) Fast personal protective equipment detection for real construction sites using deep learning approaches. Sensors 21(10), 3478.CrossRefGoogle ScholarPubMed
Wang, Z, Ying, H and Nyberg, K (2024) Graph-based change propagation across BIM design software applications. In 2024 ASCE International Conference on Computing in Civil Engineering.Google Scholar
Wang, Z, Ying, H and Sacks, R (2024) Two fundamental questions concerning BIM data representation for machine learning. In Proceedings of the 41st International Conference of CIB W78. https://itc.scix.net/paper/w78-2024-25.Google Scholar
Xia, J, Liao, W, Han, B, Zhang, S and Lu, X (2025) Intelligent co-design of shear wall and beam layouts using a graph neural network. Automation in Construction 172, 106024. https://doi.org/10.1016/j.autcon.2025.106024.Google Scholar
Ying, R, He, R, Chen, K, Eksombatchai, P, Hamilton, WL and Leskovec, J (2018) Graph convolutional neural networks for web-scale recommender systems. Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 974983. https://doi.org/10.1145/3219819.3219890.Google Scholar
Zhao, P, Fei, Y, Huang, Y, Feng, Y, Liao, W and Lu, X (2023) Design-condition-informed shear wall layout design based on graph neural networks. Advanced Engineering Informatics 58, 102190. https://doi.org/10.1016/j.aei.2023.102190.Google Scholar
Figure 0

Figure 1. Illustration of Building Information Graphs ($ G $) in partial models.

Figure 1

Figure 2. Data representation in various domains (this figure is modified based on Wang et al. (2024)).

Figure 2

Figure 3. Conceptual graph learning for building information models.

Figure 3

Table 1. Graph applications to support collaboration, coordination, and management

Figure 4

Table 2. Applications of graph generative design

Figure 5

Figure 4. Change propagation from the architecture Revit to the Tekla structure by leveraging the interdisciplinary links (adopted and modified from Wang et al. (2024)).

Figure 6

Figure 5. Examples of design constraint classes among architectural and structural slabs (adopted from Wang et al. (2023)).

Figure 7

Figure 6. Graph-based version control (adopted from Esser et al. (2022)).

Submit a response

Comments

No Comments have been published for this article.