1. Introduction
Reducing the weight of structures is an ongoing challenge of the aerospace industry, as structure weight is linked with environmental impact (Kaufmann, Zenkert, & Wennhage Reference Kaufmann, Zenkert and Wennhage2010).
Aircraft structures are traditionally built from stiffened panel assemblies (Bruhn Reference Bruhn1973). The elementary structures have specific functions, such as the ribs, spars and skins which are then assembled to create sections, such as the fuselage, the wings and the cockpit (Bruhn Reference Bruhn1973). A significant advantage of stiffened panels is their easy integration and assembly with other systems of the aircraft.
Stiffened panels are built from two sets of components, the skin and the stiffeners, see Figure 1. Functionally, the skin is sized for structural stiffness and the stiffeners are used to maintain the structure’s stability and increase local buckling resistance (Bruhn Reference Bruhn1973). The design of stiffened panels is a complex multidisciplinary activity. Its design can impact the aerodynamic, vibration, controls, systems and so forth. For structural engineers, the main challenge is the study of mechanical failures, thus ensuring the panels can withstand heavy loads throughout the lifecycle of the aircraft. Some of the mechanical failure causes are local, such as fracture, rupture, collapse, local buckling, and post-buckling instability (Megson Reference Megson2017). Other failure causes are global, for example, flutter or global buckling (Megson Reference Megson2017). State-of-the-art design methods for stiffened panels are based on handbook relations to address most mechanical failure causes (Bruhn Reference Bruhn1973). The solutions proposed in this paper are in regard to the challenge of creating panel layouts outside the traditional ones, in the hope of reducing the weight of aircraft structures (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2021).
Stiffened panels are mainly built from orthogonally placed stiffeners in the orthogrid layout (Bedair Reference Bedair2009), which is the illustrated layout of Figure 1. There are very few new layouts used in aircraft design, as they are linked to an increase in analysis, validation and certification complexity. Additionally, there is only a limited amount of empirical data available for such layouts, meaning semi-empirical calculations are not usable for their analysis. One of the reasons for the success of the orthogrid layout is its ability to maintain global stability even when local skin segments are buckling; this phenomenon is referred to as post-buckling stability (Bruhn Reference Bruhn1973; Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2019). Controlling it through sizing optimization allows the orthogrid stiffened panel weight to be reduced significantly.
Buckling containment features (BCF) and curvilinear stiffeners are some good examples of new layouts. BCFs were found to increase the strength-to-weight ratio of stiffened panels in certain regions of a commercial aircraft wing (Houston et al. Reference Houston, Quinn, Murphy and Bron2016). BCFs and curvilinear stiffeners have been proposed using the experience and intuition of engineers and researchers to improve the design of stiffened panels (Kapania, Li, & Kapoor Reference Kapania, Li and Kapoor2005; Houston et al. Reference Houston, Quinn, Murphy and Bron2017). These layouts have generated traction due to advances in manufacturing and assembly techniques such as electron beam free-forming, additive manufacturing and integrally machined stiffeners (Kapania, Li, & Kapoor Reference Kapania, Li and Kapoor2005; Caseiro Reference Caseiro2013; Mulani, Slemp, & Kapania Reference Mulani, Slemp and Kapania2013). The topological design space opened up by these new manufacturing methods is large and complex to navigate. Consequently, traditional design and optimization tools are having trouble finding innovative ways to leverage the new layouts and manufacturing techniques. Topology optimization is an example of a numerical tool that attempts to find new lighter and innovative layouts (Zhu, Zhang, & Xia Reference Zhu, Zhang and Xia2016; Aage et al. Reference Aage, Andreassen, Lazarov and Sigmund2017; Tyflopoulos et al. Reference Tyflopoulos, Tollnes, Steinert and Olsen2018).
Topology optimization has been used in multiple academic and industrial research projects (Sigmund & Maute Reference Sigmund and Maute2013; Deaton & Grandhi Reference Deaton and Grandhi2014). In current industrial practices, the solid isotropic material with penalization (SIMP) method is the most commonly used and is implemented in many commercial software (Ansys© Workbench 2020; Optistruct 2018). Older techniques include the ground-structure method (GSM) and the homogenization method (Bendsøe & Sigmund Reference Bendsøe and Sigmund2003). More recent topology optimization techniques are the level-set method (LSM) and the moving morphable component (MMC) (van Dijk et al. Reference van Dijk, Maute, Langelaar and van Keulen2013; Guo, Zhang, & Zhong Reference Guo, Zhang and Zhong2014).
In the aviation industry, SIMP has been used for the design of many aircraft components such as wings, fuselage and subsystems (Krog et al. Reference Krog, Tucker, Kemp and Boyd2004; Zhu et al. Reference Zhu, Zhang and Xia2016; Aage et al. Reference Aage, Andreassen, Lazarov and Sigmund2017). Still, no major projects have shown any improvement in the main structural component (Aage et al. Reference Aage, Andreassen, Lazarov and Sigmund2017). In our research group, we have identified buckling load factor constraints as one of the main factors that impact the viability of stiffened panel optimization for both SIMP and GSM approaches (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021). So far, our research has only been able to show that topology optimization can reach viable results, when buckling is considered, if coupled with a random restart approach that yields an improved exploration of the design space (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021). Still, random restart approaches have shown to be highly inefficient as they tend to find only the minima with the largest “valleys,” which are, at best, as efficient as the orthogrid layout (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021).
Another aspect we observed with the use of topology optimization in the industry is that the results tend to be highly complex geometries that cannot be manufactured without design simplification. However, these simplifications made by engineers tend to cut down any apparent weight gains made through topology optimization. It explains partially why it is so challenging to propagate optimized layouts throughout all the design phases. The effect of design complexity is shown well in the large-scale project carried out by the Technical University of Denmark (DTU) where the optimized material distribution sacrifices its apparent stiffness and weight advantages when interpreted to a manufacturable design (Aage et al. Reference Aage, Andreassen, Lazarov and Sigmund2017). This effect has also been observed in previous work of our research group which was carried out on ribs, pressure panels and wing components with different topology optimization algorithms (Dugré, Vadean, & Chaussée Reference Dugré, Vadean and Chaussée2016; Gamache et al. Reference Gamache, Vadean, Noirot-Nerin, Beaini and Achiche2018, Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021). Making SIMP and GSM work for stiffened panels optimization required several weeks of fine-tuning simulation and optimization parameters towards results that can be ultimately transferred to detailed design and manufacturing (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021). This tuning is necessary, as not all load cases and mechanical failure causes can be included as topology optimization constraints, due to generation capabilities and limited computing power. This leads to a classic optimization issue, where an optimization solver will optimize out any considerations that cannot be included in the optimization formulation. This dichotomy between design and optimization makes optimization difficult to use for the design of structures, and is especially true for the more complex layout optimization problem.
In summary, our previous work has not shown topology optimization to be ready for larger use cases, especially due to difficult tuning requirements, high computing costs and solutions’ complexity (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021).
Current topology optimization algorithms and processes, when applied for layout optimization, have yet to efficiently find layouts lighter than conventional layouts. Considering the limitations discussed above, we have developed a new heuristic two-level optimization process specifically for stiffened panel layout optimization. There have been other works on two-level optimization using topology optimization, but none on stiffened panel layout or using complexity (Chirehdast & Papalambros Reference Chirehdast and Papalambros1992; Walbrun, Witzgall, & Wartzack Reference Walbrun, Witzgall and Wartzack2019).
Topology optimization is an important computational tool that reassembles traditional design searches (Hay et al. Reference Hay, Duffy, McTeague, Pidgeon, Vuletic and Grealy2017) and as such our new tool will be based on a design-centric approach. Our development approach is based on the generative design framework discussed in the first author’s thesis (Gamache Reference Gamache2021). This framework distinguishes three properties of generative systems along three axes: generation, representation and navigation. These properties are defined as follows:
Generation refers to the way the system infers the architecture of the solution from the description of the topological variables. Generative systems can range from implicit to explicit generative variables, where implicit variables describe a physical property of a specific volume while an explicit variable would describe the relationship between discrete components, see Figure 2.
Representation refers to the degree of fidelity used for the performance evaluation of the generative system. A gradient exists between low-fidelity to high-fidelity simulation, as low-fidelity is less computationally expensive, with the compromise of less complicated models. Note that low-fidelity simulations can still be highly accurate, but require well-understood simulation hypotheses.
Navigation refers to the axis between exploiting a small but promising region of the search space, and exploring through a larger region for possible better optimum (Gamache Reference Gamache2021). In optimization vocabulary, it refers to whether the process is looking for local or global optima.
The advantages of using the generative design framework are its effectiveness to identify trends and its common usage in the topology optimization community. First, we identified that the new research topics are moving from implicit (e.g., SIMP) (Bendsøe & Sigmund Reference Bendsøe and Sigmund2003) towards a more explicit generation (e.g., moving morphable components) (Guo et al. Reference Guo, Zhang and Zhong2014). Second, adding physics considerations such as stress and buckling, has been an ongoing issue since it translates into nonlinear constraints (Ferrari & Sigmund Reference Ferrari and Sigmund2019). In turn, the nonlinear constraints induce feasible spaces that are more difficult to navigate, requiring even more computing power. Finally, we observe a major difference between academic work and industrial usage of topology optimization, where in academia topology optimization is presented as a tool to find an optimum in a convex design space, thus focusing on exploitation, while in the industry topology optimization is rather used to diversify solutions, thus focusing on exploration.
These observations led to the conclusion that the way topology optimization was carried out differently in academia and in industry can explain, in part, why its use for aircraft structure innovation has not been successful yet. The challenges of topology and layout optimization for aircraft structure are explored further in the next section.
2. Using complexity to drive the topological design space navigation
Following the trends identified in the generative design framework in stiffened panel topology optimization, we created a new process that addresses the layout optimization challenge. This process integrates design considerations during optimization and focuses on exploring simple, viable solutions while limiting the computing time. The key to efficiency is the new complexity objective that drives navigation and allows adding complexity to the design (layout) only if it is necessary. We named the new process after this key feature: complexity-driven layout exploration for aircraft structure (CD-LEAS).
Integrating a nonlinear constraint, such as buckling, requires a simulation model with more elements than ones aiming only for compliance but does not need an increase in design variables. In SIMP, the simulation model and optimization models use the same elements and as such, increasing representation fidelity also increases optimization cost with no added benefits. Consequently, to reduce computation cost, it is crucial to decouple the simulation and the optimization models. Decoupling a process depends on the type of generation used, and is made easier with more explicit approaches such as MMC (Guo et al. Reference Guo, Zhang and Zhong2014).
Explicit generation creates a navigation challenge as the design space becomes disconnected. Therefore a disconnected space cannot be easily navigated by gradient-based optimization. To search in this design space, a heuristic optimization approach is better suited. However, it is known to be less efficient than gradient-based algorithms. This low efficiency is alleviated in CD-LEAS by using a novel complexity measure as an optimization objective, as it limits the search to low-complexity solutions. Intuitively, using complexity as an objective and starting with a low-complexity layout emulates the thinking process of an experienced engineer, which usually starts with simple solutions, and increases complexity only if it brings any benefits towards the design objective. In other words, the complexity objective modifies the search pattern to easily explore first the smaller low-complexity area before moving on to the larger high-complexity area of the design space. Moreover, using an explicit design space has the advantage of making significant topological changes between each optimization iterations, whereas gradient-based solutions require a simulation of multiple solutions with the same layout, but with slightly different design parameters. These search patterns reduce significantly the number of inner-loop optimization runs. Using the complexity measure to infer direction to the navigation is the key element of the CD-LEAS process.
The CD-LEAS process combines know-how from many disciplines to achieve an effective design exploration, see Figure 3 for an illustration of the process. CD-LEAS can be described as a two-level optimization process, where layout/topology is optimized in the outer loop and sizing is optimized in the inner loop. The layout optimization formulation is detailed in the navigation section and the sizing optimization is in the representation section of this paper. Another advantage of the CD-LEAS process is that the explicit generation allows for an easy integration of the two optimization loops, rather than an approximation between the two loops as done in previous research work (Bremicker et al. Reference Bremicker, Chirehdast, Kikuchi and Papalambros1991; Chirehdast & Papalambros Reference Chirehdast and Papalambros1992; Gamache et al. Reference Gamache, Vadean, Noirot-Nerin, Beaini and Achiche2018).
For the generation, CD-LEAS uses a graph-grammar that can create many reinforcement layouts from actions and rules. Graph-grammars allow for a quick explicit generation of layouts that also ensures significant topological changes for each iteration. In this work, the generation is implemented in Matlab (MATLAB 2020).
As the generation is carried out using an explicit method, the choice of the fidelity of the representation remains flexible. The limitation here is the software and hardware available to the user. As such, any fidelity level can be used, from linear static to nonlinear geometric analysis. For the implementation discussed in this paper and the use case described, we opted for linear models for both compliance and buckling to keep computation time reasonable but simulation accurate. We developed an automation script in HyperMesh (HyperWorks 2018) to create a finite-element model (FEM) for each proposed topology. Furthermore, each layout is sized with an inner optimization loop and will be discussed in more detail in the representation section. HyperMesh and Optistruct from Altair (HyperWorks 2018) are used to create, evaluate and size each model. The automation is done via the TCL scripting capabilities of HyperMesh (HyperWorks 2018).
Finally, for navigation, CD-LEAS uses a relative functional complexity measure as one of the objectives to drive the design space exploration. This measure integration has two advantages: first, it ensures that layouts can be more easily transferred into detailed design and second, that optimization seeks complex solutions only if it benefits the objective function. To navigate the disconnected topology design space, we decided for a stochastic graph search algorithm, namely the basic Burst algorithm (Königseder & Shea Reference Königseder and Shea2014). The navigation is also implemented in Matlab (MATLAB 2020).
As for the structure of this paper, we first discuss the source and usage of the proposed functional complexity measure. We then develop how the generic complexity measure is applied for the specific case of layout optimization of stiffened panels. The measure is presented in parallel with the description of the inner sizing optimization loop. From there, the discussion focuses on the graph-grammar generation process. Our implementation of the basic Burst algorithm is explained with the adaptation necessary for complexity-driven navigation. Finally, two case studies of the common usage of stiffened panels are proposed to showcase the efficiency of CD-LEAS. The first focuses on a rectangular pressure bulkhead and the second on a highly compressed, buckling-sensitive panel.
3. Complexity measures for optimization
In the topology optimization community, some published research has implemented complexity constraints to topology optimization. In density-based topology optimization, size filters are considered complexity limits (Sigmund Reference Sigmund2007). In GSM, the number of nodes, edges and connections are used to measure and constrain geometric complexity (Torii, Lopez, & Miguel Reference Torii, Lopez and Miguel2016). In LSM, the complexity is calculated depending on the number of basis functions (Zhang et al. Reference Zhang, Liu, Wei, Zhu and Guo2017). Finally, for MMC, complexity is measured by the number of effective components (Zhang et al. Reference Zhang, Zhou, Zhu and Guo2017). In summary, in the topology optimization community, complexity is only measured with regard to the geometry of the solution and not its overall design challenge. In this work, design complexity refers to the difficulty of subsequent design phases. Using design complexity to drive an automated design and optimization process has not been done before, to the best knowledge of the authors. As discussed earlier, the nature of the complexity measure for CD-LEAS is not as important as the fact it exists to shape the design space towards something that is manageable by optimization in a reasonable time.
From the existing topology complexity measures, we found none that answered our need to translate design complexity as an optimization objective. There is already existing knowledge on this aspect in the axiomatic design community, for which the authors were already familiar with (Suh Reference Suh2005). As such, we leveraged existing know-how and saw a clear path from the qualitative approach towards a quantitative measure that we could implement with stiffened panels. Measuring design complexity is an ongoing area of research, with perspectives from many different disciplines (Min, Suh, & Hölttä-Otto Reference Min, Suh and Hölttä-Otto2015; Mohebbi, Achiche, & Baron Reference Mohebbi, Achiche and Baron2018; Sinha, Suh, & de Weck Reference Sinha, Suh and de Weck2018; Chouinard, Achiche, & Baron Reference Chouinard, Achiche and Baron2019; Bjarklev et al. Reference Bjarklev, Eifler, Mortensen, Linnebjerg and Ebro2020). In this work, we focus on the definitions used in axiomatic design for conceptual design. As such, complexity is not only related to geometry, but it also becomes dependent on the specific design problem. Qualitative aspects such as a measure of interaction between discrete components are often used in complexity measurement literature. This concept is strongly linked to axiomatic design through the first axiom of Independence. Measuring for design parameters interaction is a good evaluation of complexity, as unpredictable interactions will push the design process towards an inefficient round of trial-and-error to identify a successful solution. As such, complexity is not only related to geometry, but it is also dependent on the specific design problem. The measure we propose is relative and allows for an easier synthesis of multiple solutions with respect to a specific problem.
4. Complexity measure with axiomatic design
Axiomatic design is born out of a need to create a systematic approach for the design process (Suh Reference Suh2005). Axiomatic design provides a framework to understand the effect of design on the quality of the product. In other words, it makes sure that you do the right things and that you do things right (Farid & Suh Reference Farid and Suh2016). One of the main aspects of axiomatic design is the separation of the design activity into four domains, customer attributes (CA), functional requirements (FR), design parameters (DP) and process variables (PV) (Suh Reference Suh1990).
For CD-LEAS’ complexity measure, we focus on the relationship between the FRs and DPs, which is characterized and given by the design matrix ( $ \mathbf{A} $ ). The design matrix can help identify the types of information that are present in the current layout (see Eq. (1)). The design matrix may be built in two different ways, either as a set of linear equations or using the differential relationship of each FRs with respect to each DPs (see Eq. (2)). In this work, we use the differential design matrix in order to leverage information from optimization:
In addition to the domains, there are two identified axioms that stand as the basis of axiomatic design: the Independence and the Information axioms (Suh Reference Suh2005). The first axiom stipulates that a good design maintains independence between the FRs. The second axiom puts emphasis on the need to minimize the Information content of the design. The Information content is defined as the probabilistic relation related to the number of combinations that leads to uncontrolled FRs with respect to the DPs (Suh Reference Suh2005). Said otherwise, the Information content of a design reflects the probability of DPs not satisfying every FRs.
Using these axioms as starting points, Suh has defined complexity as the difficulty to fulfill these two axioms in a single design (Suh Reference Suh2005). Further work by Puik & Ceglarek (Reference Puik and Ceglarek2016) has extended this idea, by developing a design complexity theory that uses Information content as its focus point. Their theory separates design Information into different categories, providing a framework to understand complexity:
Traditionally, the visualization of the design matrix allows engineers to categorize solutions between coupled (challenging), decoupled (acceptable) and uncoupled (ideal) solutions (Suh Reference Suh2005). See Eqs. (3), (4) and (5) for the possible coupling levels observable in the design matrix. Uncoupled solutions are the easiest solutions to implement, as no coupling between DPs can perturb the FRs of the design thus making design easy independently of the design decisions order.
Uncoupled matrix:
Decoupled matrix:
Coupled matrix:
In axiomatic design literature, there are two measures based on the design matrix for assessing the coupling of the design matrix: re-angularity and semi-angularity (Suh Reference Suh1990). Re-angularity measures the coupling of each FRs, with respect to all DPs (Suh Reference Suh1990). In a similar fashion, semi-angularity can detect if a system is decoupled or uncoupled (Suh Reference Suh1990). Both of these measures are not meant to measure the relative complexity of a given system, it only classifies the design matrix in one of the three coupling categories. Still, we keep in mind what they are trying to measure: the coupling of FRs, that is, the recognized Information content in Puik’s framework.
5. Complexity measure for complexity-driven exploration
The present work uses the information complexity framework proposed by Puik and Ceglarek, which was discussed in the last section (Puik & Ceglarek Reference Puik and Ceglarek2016). Note that our goal is not to directly measure the information content of each design, but to develop indirect approximations that correlate with the Information content. Unrecognized information cannot be identified using only analytical data, but the three other types (recognized, superfluous and axiomatic) can be approximated to reduce unnecessary complexity.
In this section, the three approximations are introduced, with the symbol $ \Psi $ representing a complexity approximation. These approximations will be aggregated using a simple Euclidean norm to approximate the useful (or total) Information content, see Eq. (6).
where $ {\Psi}_U $ is the useful Information content, $ {\Psi}_A $ is the axiomatic Information content, $ {\Psi}_R $ is the recognized Information content and $ {\Psi}_S $ is the superfluous Information content.
Let us start by approximating the recognized Information content, which is the coupling of the FRs with respect to DPs. Said in mathematical terms, we seek to measure the collinearity of the FR vectors in the design matrix. A good tool to measure the collinearity of vectors is the matrix conditioning number. In this work, we start from a generic definition (see Eq. (7); Trefethen & Bau Reference Trefethen and Bau1997). This conditioning number uses the singular value decomposition (SVD) to measure if any collinearity in the design matrix. For our implementation, we can use the function cond() natively provided in Matlab (MATLAB 2020).
However, this definition only uses the extreme (min and max) measures of the matrix and as such can only identify if there is any collinearity or not at all. It is a very useful measure of the complexity of computing the inverse of an arbitrary matrix, but not a good measure of recognized Information. Consequently, as a recognized Information measure, we propose a normalized sum of all SVD components:
where $ {\Psi}_R $ is the recognized complexity, the $ SVD $ operation returns the singular value decomposition vector and the $ \mathit{\operatorname{rank}} $ operation is the size of the design matrix. This measure of recognized Information content is relative. When $ {\Psi}_R $ goes to zero, there is no collinearity, as all the SVD elements are unity. If there are multiple collinearities detected, then the measure goes up to one.
Another aspect that is useful to measure is the axiomatic Information, which measures how difficult it is to realize the functions with a given set of DPs. It is an indirect measure of the robustness of the DPs with respect to the FRs. To approximate axiomatic Information, we use the condition number of a nonlinear function (Trefethen & Bau Reference Trefethen and Bau1997), as defined in Eq. (9). The condition number of a nonlinear function reflects the sensitivity of the FRs with respect to the DPs (Trefethen & Bau Reference Trefethen and Bau1997).
where $ {\Psi}_A $ is the axiomatic complexity, $ J(x) $ is the Jacobian matrix of the sizing problem at the local optimum, $ f(x) $ is the performance at the local optimum and $ x $ is the DPs at the local optimum. The Euclidean norm was used for this equation, but any matrix and vector norm can be used. For our application, as we use the differential notation of the design matrix $ \mathbf{A} $ , we can simply replace the design variable as the Jacobian $ J(x) $ in the axiomatic complexity equation. From our experiences with stiffened panels, $ {\Psi}_A $ usually varies between 0 and 5. When $ {\Psi}_A $ goes closer to zero, it means that big variations of DPs have a low impact on FRs. As it gets higher, the impact of DPs increases on FRs, making the design more complicated to manage.
Finally, superfluous Information can arise when some DPs have no impact on the FRs, they just make optimization more difficult. As such, we define the amount of superfluous information using an arbitrary threshold (we use 10%) on the norm of the vector of the influence of each DPs, as shown in Eq. (10).
where $ {\mathbf{A}}_{reduced} $ is the reduced design matrix, from which DP vectors with a norm below 10% of the maximum vector norm of $ A $ are removed.
Unrecognized Information is not captured in the design matrix, and as such we are not able to measure it. Still, solutions with low complexity, with a low recognized, axiomatic and superfluous information, will have an easier time managing the added Information that can arise during testing, production and maintenance.
In summary, each of these Information content approximations can be used to approximate a relative complexity measure. This measure of complexity could potentially be used for any design problem where the design matrix can be computed in its differential form. In the next section, we introduce how the complexity is measured specifically for stiffened panels as well as a description of the inner optimization loop.
6. Representation: Complexity, objective and constraints evaluation for stiffened panels
For CD-LEAS, we need to be able to measure performance and complexity automatically for any layouts created. In order to do so, we present here how we deal with the FRs, constraints and performance of stiffened panels using a FEM-based inner loop sizing optimization. Stiffened panels are composed of multiple subcomponents; the main component is the skin to which stiffeners are added in order to increase strength (Megson Reference Megson2017). The stiffeners can be built in different ways, but generally have a pad-up section, a web that is perpendicular to the skin and a free flange to increase the buckling resistance of the stiffener (see Figure 1). The panels are periodically attached to bigger assembly components such as the spars and ribs (Megson Reference Megson2017).
As the stiffened panels are built using thin sheet metal or laminated composite components, buckling stability, imperfections and out-of-plane forces are the main concerns for their design (Bruhn Reference Bruhn1973). The most used design method for stiffened panels is based on handbooks (Bruhn Reference Bruhn1973; Megson Reference Megson2017) that contain important relations and equations for the orthogrid layout. For example, analytical and semi-empirical equations exist to predict skin and Euler buckling, as well as Euler-Johnson’s post-buckling collapse (Bruhn Reference Bruhn1973, Megson Reference Megson2017).
In this work, a more flexible analysis approach is necessary to model any layouts. FEM is a flexible and accurate approach, at the trade-off of more computing power than handbook methods. The buckling analysis is done via eigenvalue analysis. Post-buckling is out of the scope of the current implementation of CD-LEAS, but it could be implemented in a future revision.
To help with the synthesis and comparison of the different layouts, an inner loop sizing optimization is included in the performance evaluation. For the inner loop, we use Optistruct, which includes a family of efficient gradient-based solvers. The sequential quadratic programming (SQP) algorithm is used for the sizing problem.
From experience, we have observed a good convergence towards the same minima from different initial values, indicating a relative stability of the optimization. Evaluating the design space convexity is out of the scope of this work. Of course, more work and computing power could be leveraged in future studies to ensure global optimality of the inner-loop optimization problem.
The objective of the sizing optimization can vary depending on the use case. It could be a weight minimization with regard to nonlinear constraints, or a simple compliance minimization given a weight constraint. The exact formulation of the sizing optimization depends on the case study and is discussed in their respective sections.
At the optimum obtained from the sizing optimization, it is possible to measure the layout’s complexity. At this point, the distinction between the optimization and design problem is central, as the optimization problem may be formulated differently, as to approximate the design problem. In CD-LEAS, thickness variables and the compliance objective from the optimization are selected as the simplified DPs and FRs of the design problem, as described in Eq. (11). To note, this optimization formation is different than the inner-loop optimization problem, and is only used to compute sensitivity, which is used to approximate the design matrix $ {\mathbf{A}}_{\mathbf{SP}} $ .
where $ \mathbf{t} $ is the thickness of each segment of a given layout, $ {C}_i $ is the compliance of each segment, $ i $ is the segment id and $ m $ is the number of segments.
Compliance, computed as the deformation energy, is most often used in topology optimization as the main optimization objective (Bendsøe & Sigmund Reference Bendsøe and Sigmund2003). It allows for a fast convergence towards not only stiff, but coherent material distribution (Bendsøe & Sigmund Reference Bendsøe and Sigmund2003). In topology optimization, a coherent material distribution ensures structural stability (e.g., no unexpected possible displacement) and connectivity of boundary conditions. As such, compliance is the best quantifiable mean to represent the FR of structures.
While topology optimization commonly measures the compliance of the whole structure by measuring the total strain energy, in CD-LEAS we are instead interested in the local compliance of discrete skin and stiffener segments. As such, in this work we define the FRs of the stiffened panel as the stiffness of each component, that is each skin and stiffener sections (measured as their local compliance). The DPs of a given layout are defined as the thickness of each subcomponents.
Any other responses, such as weight, stress, buckling or fatigue, can be considered as design constraints and consequently, they are not required to adhere to the design axioms (Suh Reference Suh2005). Therefore, it is important to point out the difference between design constraints and optimization constraints, as well as the nuance between FRs and optimization objectives. The optimization objectives and constraints are to be defined from the FRs, DPs, and design constraints while the optimization is built to answer a specific design question. For example, an optimization problem could answer the question: “What is the minimal weight this layout could achieve, with respect to a stress constraint?”. For this question, the FRs and DPs remain stiffness and thicknesses, but the optimization objectives and constraints are respectively weight and stress. As such, a well-defined optimization problem can use design constraints as optimization objectives, while the design FRs are used as optimization constraints. They are related, but not causal. More discussion on this in the case study section.
Using local compliance measures as the FRs allows the process to build the design matrix from the sensitivity analysis of the sizing optimization, with regard to the thicknesses and compliance of each subcomponent, see Eq. (12). Emphasis here, the optimization sensitivity is an approximation of the design problem and does not need to be the same as the inner-loop optimization problem.
where $ {\mathbf{A}}_{\mathbf{SP}} $ is the design matrix of a given stiffened panel, $ {C}_i $ denotes the compliance of each subcomponent and $ {t}_i $ is their thickness.
With this design matrix for stiffened panels, it is possible to implement the complexity measure of discussed in the previous section. The updated Eqs. (13), (15), (14) update the design matrices from $ \mathbf{A} $ to $ {\mathbf{A}}_{\mathbf{SP}} $ . We also updated the FR and DPs from generic to applied; $ f\left(\mathbf{x}\right) $ and $ \mathbf{x} $ to $ \mathbf{C}\left(\mathbf{t}\right) $ and $ \mathbf{t} $ .
Until now, we have described a general functional complexity measure and then how it can be used for a sizing optimization problem of a stiffened panel with any layout. In the next section, we describe a graph-grammar implementation that can create any layout with straight stiffeners from simple design rules.
7. Generation: graph-grammar for stiffened panels
In generative design, graph-grammars are used to generate design concepts from explicit design actions and rules that affect each component and their relationships. There are multiple types of graph-grammars and in this work, we have chosen to work on a set grid with a simple graph-grammar composed of only one possible type of action and a few rules. We use a fixed grid for two reasons. First, it allows for an easy implementation of a “bar code,” which is a unique ordered string that describes the topology of the layout. The bar code ensures that the navigation algorithm does not investigate the same solution twice, via different decisions path. Secondly, fixed-grid solutions are much easier to work with for a reinforcement learning algorithm, which will be studied in the future in our research group.
Using this grid, the only available action for our grammar is named “CreateStiffener,” which has to respect certain rules before being applied. This action creates an edge and connects two nodes on an existing grid, activating each node it passes through. See Figure 4 for an example of actions with a 5 × 5 grid. To define the rules of our implementation, each node of the fixed grid has a specific status. They are either “active,” “offset,” “side” or “inactive.” “Active” and “Inactive” refer to nodes with or without a connected edge respectively “Side” nodes are a special case of “active” nodes, that is, they are always considered active but will only allow the action “CreateStiffener” between nodes of different sides. “Offset” nodes are semi-active nodes, in the sense that if a stiffener passes in their area of influence, the “offset” node becomes available for connection and their position is modified to be at the crossing of the new stiffener. Implement the “offset” nodes allows the identification of a graph independently of the sequence of rules required to get there.
With this single action and rules, the graph-grammar can create a graph representation of any piecewise linear stiffening layout. From this graph representation, we can convert the position of nodes and edges connectivity into a parametric file that is passed to a TCL script that builds a FEM model of the given layout using HyperMesh automation capabilities (HyperWorks 2018). The TCL script also calls Optistruct for the inner-loop optimization problem. Also, as the node positions are described on a grid from 0 to 1 in 2D, the TCL script can easily extrude the stiffeners with respect to the normal of any planar or curved plane using a bilinear projection.
This graph-grammar is more explicit than GSM topology optimization for a simple reason: the decisions are not made for each single edge. Said otherwise, the layout is not fixed, only the grid. The process is able to create stiffeners, by connecting edges without preexisting connections. This approach ensures that significant topological changes are made at each decision. Another advantage of the graph-grammar is that the description of the layout is independent of the sequence to get to the solution. In the future, more complex actions could be built, such as “create grid”, “copy pattern”, “mirror pattern” or even introduce curvilinear stiffeners.
Still, one limitation is to be noted. As the grid is fixed, where more than one stiffener passes near an “offset” node, the offset is removed and all the edges connect at the fixed node. This tweak ensures that a given layout, with a given bar code, reflects the same structure independently of the order of previous decisions. So, for a more “precise” description, a finer grid of nodes can be used, although it significantly increases the design space. For this current study, a limit to 10 × 10 grids is set, which is still more than enough to find interesting solutions for simple stiffened panels. We also implemented a symmetry option, which helps reduce the size of the problem.
The graph-grammar is implemented as a class in Matlab. This class is capable of returning a list of possible actions that respond to the design rules, and in the next section, we describe the algorithm used to choose which actions to take in this list.
8. Navigation: complexity-driven basic burst-algorithm
The list of possible actions for any layouts is large. Therefore, the stochastic search process is about searching effectively through a decision tree with controlled random decisions that reduce the total amount of required performance evaluations. Commonly used stochastic search algorithms are the greedy search, random search, random walk, evolutionary algorithms, etc. (Hoos & Stützle Reference Hoos and Stützle2004). The choice of a search algorithm is driven by the need to balance randomized and goal-directed search; decide which is more important in terms of exploration versus exploitation (Hoos & Stützle Reference Hoos and Stützle2004). For our case of multiobjective graph-grammar navigation, we found that the Burst algorithm provided simple parameters that can be leveraged to push navigation towards exploration rather than exploitation (Königseder & Shea Reference Königseder and Shea2014). A basic and advanced version of the Burst algorithm are proposed in Königseder & Shea (Reference Königseder and Shea2014). For the sake of simplicity, this work uses our own implementation of the basic Burst algorithm. A possible future improvement for CD-LEAS could be to implement the advanced Burst algorithm or any more efficient search algorithm.
Our implementation of the basic Burst algorithm is illustrated in Figure 5. The algorithm is initialized with any number of initial layouts that are placed in the archive. Throughout the search, the archive will contain all layouts and their properties. At each iteration, a Pareto Front is built from the layouts in the archive. In our implementation, we used the script available from the MathWorks file exchange (Tom Reference Tom2021). From the Pareto front, we randomly choose $ N $ layouts that are created, updated and evaluated in parallel. $ N $ can easily be scaled to the number of cores available in the user’s computer or server. We added a condition to ensure that each layout of the Pareto front is chosen at least once for improvements at each iteration. As we found the Pareto front to be too small at the beginning of the search, a relaxation mechanism was used; if the number of layouts on the Pareto front is smaller than $ N\ast r $ , where $ r $ is the relaxation parameter, we redo the Pareto evaluation on the remaining nondominant layouts. This step is repeated until there are at least $ N\ast r $ layouts in the active search set. In our test cases, it was found that using $ r=3 $ yields a good balance between exploration and exploitation. In short, $ N $ and $ r $ are search parameters that can be used to tune exploration versus exploitation. If we increase the values of $ N $ , there are going to be more decisions made from a single set of layouts without looking at performance. Navigating randomly in short bursts can increase exploration and diversity, at the cost of more evaluations that might lead to uninteresting layouts. At the opposite, small values of $ N $ will make the algorithm behave more like a greedy search, thus improving the exploitation at the cost of exploration. The only other search parameter is the stopping criteria $ {Archive}_{max} $ which is simply the maximum number of created layouts.
To summarize this method section, CD-LEAS introduces a balance of generation, representation, and navigation compared to current topology optimization methods. This balance is closer to what is used in industry, where exploration and simplicity are key to finding new sound and manufacturable solutions. The relative complexity measure is used to focus on exploration as well as to limit the search within simple layouts. The graph-grammar allows for the generation of solutions that is decoupled from the performance evaluation. And finally, our implementation of the basic Burst algorithm is used to explore the design space, with a good balance of exploration versus exploitation.
The next section will demonstrate the capabilities of CD-LEAS on some challenging industrial topology optimization problems that our research team encountered previously (Dugré et al. Reference Dugré, Vadean and Chaussée2016; Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020, Reference Gamache, Vadean, Dodane and Achiche2021).
9. Case studies of CD-LEAS
In this section, we propose two different case studies of common loading conditions of stiffened panels: a pressurized bulkhead and a highly compressed wing section panel of a wing section.
The only elements that change between the examples are the boundary and load conditions. Even then, the case study has very different design spaces for which CD-LEAS can find adequate results. Furthermore, a comparison with the SIMP method for the pressurized bulkhead and handbook optimization for the compressed wing section are presented.
10. Pressure bulkhead with out-of-plane loading
The pressure bulkhead is a known difficult challenge for topology optimization (Dugré et al. Reference Dugré, Vadean and Chaussée2016; Warwick, Mechefske, & Kim Reference Warwick, Mechefske and Kim2019). In our previous work, several shortcomings of SIMP were found in this case study, finding that the interpretation required an extensive study of the layouts proposed by topology optimization (Dugré et al. Reference Dugré, Vadean and Chaussée2016).
11. Description of the pressure bulkhead case study
In aircraft, bulkheads are at the front and back of fuselages to keep a stable pressure in the cabin during flight operations. It is usually round, but was kept as a square panel in this study for the sake of simplicity.
The plate is flat on the $ X $ – $ Y $ plane, modeled using Shell (PSHELL in Optistruct) elements with quadratic interpolation (CQUAD8 in Optistruct). These elements are used to model accurately the rotation and displacement, as they have six degrees of freedom, but with reduced computation cost relative to 3D elements. This modeling approach is accurate as the modeled structure is made of thin components. The boundary conditions are simply supported ( $ {U}_Z=0 $ ) on all sides and pressure is applied evenly on all the plate as a distributed normal force $ {P}_z $ , see Figure 6. Aluminum 7075 is used and its properties are found in Table 2. The graph-grammar actions and rules are the same as presented in the previous sections.
CD-LEAS can be discussed as a two-level optimization scheme, with the outer loop focusing on the layout/topology and the inner loop with sizing the components. For the pressure bulkhead case study, we are seeking an optimal placement of stiffeners with respect to a weight constraint. This formulation is selected to be able to compare easily with topology optimization. The weight constraint is only applied in the inner loop, and layouts that cannot achieve the weight constraint are eliminated from the outer loop. This makes the implementation of the layout optimization easier, as the single objective gradient descent of the inner-loop can deal with optimization constraints more easily than the heuristic multiobjective outer loop. The topology multiobjective outer-loop optimization formulation is given in Eq. (17).
where $ \mathbf{D} $ is the layout graph (topological variables), $ {C}_i $ is the compliance of each subcomponent, $ {\Psi}_U $ is the useful complexity and $ n $ is the number of subcomponents of a given $ \mathbf{D} $ layout.
For the inner loop sizing optimization of this test case, the compliance is minimized with respect to a weight constraint, see Eq. (18).
where $ \mathbf{t} $ is the thickness of each subcomponent, $ {C}_i $ is the compliance of each subcomponent, $ {w}_i\left(\mathbf{t}\right) $ is the weight of each subcomponent and $ n $ is the number of subcomponents of a given $ \mathbf{D} $ layout.
The weight constraint is set to 5.0 lb. and the value of the pressure is arbitrarily set to 10 psi. An arbitrary pressure value does not have any impact on the layout in the case of compliance minimization, as compliance is directly proportional to the applied loads. What is important for the compliance value is the direction and location of the load.
As a layout reference, we ran the SIMP method with Optistruct (Optistruct 2018) and converted the results to an explicit stiffened panel, see Figure 7. SIMP was run with standard parameters, penalty factor is set to 2, volume constraints to 20%, and shell thickness to 1 inch. The results for the SIMP results can be found in Table 1. Finally, the Burst and graph-grammar parameters are defined as follows for the pressure bulkhead case.
These parameters offer a good balance between exploration and exploitation and computation time is 25 minutes with an AMD Ryzen 7 3700X @4.0GHz, running the eight layout evaluations in parallel on the eight cores.
12. CD-LEAS results for the pressure bulkhead
The results of CD-LEAS for the pressure bulkhead are presented in different ways. First, the layouts of the final Pareto set are presented in Figure 8. Then, the performance of the entire archive is presented in Figure 9a. Finally, the convergence plot of the compliance is shown in Figure 9b. It is possible to see that convergence is obtained quickly, in fewer than 20 iterations.
Of course, using a stochastic search approach is highly dependent on the random choices the algorithm makes during exploration. It is possible to see in Figure 8 that the obtained layouts share some similar patterns. We have rerun the search with the exact same parameters and initial layout, for which the solutions of the Pareto front are illustrated in Figure 10.
In both cases, CD-LEAS is able to find solutions that are both stiffer and less complex than the results obtained with SIMP. The best CD-LEAS result has a compliance of 37, a reduction of 35% in comparison to the layout found with SIMP. However, it comes at the cost of a more complex solution. If we chose a layout similar in complexity at $ {\Psi}_U\approx 1.7 $ , the layout of Figure 8 with $ {\Psi}_U\approx 1.62 $ (third row, second column) has a compliance of 41.2, still a reduction of 28% in compliance with a similar relative complexity. For this interpretation, it is important to consider that useful complexity is approximated by the Euclidean norm of three factors: recognized Information (defined by the coupling of stiffeners sensitivity), axiomatic Information (which relates to the sensitivity of the solution to the thickness of stiffeners), and superfluous Information (which concerns whether all stiffeners are meaningful or necessary). In the context of this specific comparison, it means that even though the two layouts may look very different, they present a similar design challenge.
Furthermore, the Pareto set of the pressure bulkhead problem shows a clear compromise between complexity and compliance. For the layout design of stiffened panels, obtaining components with high stiffness is regularly not sufficient for design, as there are generally other design constraints, which can be structural. Often, constraints come from other engineering disciplines involved in the design process such as manufacturing, systems or aerodynamics. Selecting only stiffness constraints will make for difficult post-processing of the results. As an example, an optimized material distribution for a wing box would not leave any place for fuel tanks. The Pareto set can empower designers to easily choose an adequate level of compliance and complexity that should allow for a much easier subsequent detailed design phase. Our industrial partners have also found it particularly useful in a multidisciplinary approach, as it can reduce the amount of interaction with other disciplines. This trade-off can be translated as a clear hierarchy of the primary and secondary stiffening components, with components of simple solutions offering the most effective compromise of complexity stiffness. The results also show a clear advantage with respect to exploration when compared to gradient-based topology optimization. Furthermore, results obtained with SIMP are often organic and complex, which creates design challenges down the road, especially with respect to unrecognized information.
In summary, CD-LEAS can rapidly generate simple and efficient solutions, even offering better results than SIMP for the case of the pressure bulkhead. A major advantage of CD-LEAS against topology optimization, such as SIMP and LSM, is the fact that the stiffeners are explicitly generated, ensuring a proper representation of the stiffened panel which does not require any further interpretation. Besides, using the Pareto front as the starting point of each iteration allows the Burst algorithm not to get stuck in the design space of complex solutions, as the design space is smaller in the simple space than in the complex space.
The next section discusses the compressed panel case study as an in-plane loading.
13. Compressed stiffened panel, with in-plane loading
The design of compressed stiffened panel is also an active challenge of topology optimization, mainly due to convergence difficulty created by buckling load factor constraints (Layachi, Xu, & Bennaceur Reference Layachi, Xu and Bennaceur2017; Chu et al. Reference Chu, Townsend, Featherston and Kim2020; Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2021). These panels are found on the outer skins of the wing, mainly on the upper skin. The bending action of the lift and gust forces on the wing creates high compressive loads on the panels.
In the industry, these panels are currently designed and optimized using a handbook method (Bruhn Reference Bruhn1973; Megson Reference Megson2017). Relations and equations found in these handbooks are only for the orthogrid layout and are based on analytical models and empirical data to model the collapse of the panels in terms of buckling, crippling and post-buckling. The handbook design method is very efficient for large-scale problems and can easily be automated once the layout is fixed. As such, in this section, we will compare the results of CD-LEAS with the results of the design with the handbook method.
14. Description of the compressed stiffened panel case study
We propose in this section two different optimization problems, for the same design use case. The two optimization problem is used to showcase the effect of nonlinear optimization constraints on the solutions. In the first problem, we use the same optimization formulation as for the pressure panel, that is compliance and complexity minimization as given in Eq. (17) for the outer loop and compliance minimization with weight constraints for the inner loop from Eq. (18). Secondly, we introduce the buckling resistance optimization where we minimize weight and complexity, see Eq. (19), and with the inner-loop working with a buckling resistance constraint, shown in Eq. (20). This second formulation is selected as answering the question of the minimal weight of panels is the key. Here again, the buckling load factor constraint is only applied during the inner-loop optimization to ease the implementation of the outer-loop multiobjective problem. To note, in the buckling optimization case, compliance is only used for the complexity measure, which is acceptable as the design problem can be different from the optimization problem.
where $ \mathbf{D} $ is the layout graph (topological variables), $ {w}_i $ is the weight of each subcomponent and $ {\Psi}_U $ is the useful complexity.
where $ \mathbf{t} $ is the thickness of each subcomponent, $ {w}_i $ is the weight of each subcomponent and $ {\lambda}_1 $ the first buckling load of the structure.
The same graph-grammar actions and rules that were presented previously are used, with the only possible action is “CreateStiffener.” Here also, the plate is flat on the $ X $ – $ Y $ plane and the boundary conditions are simply supported ( $ {U}_z=0 $ ), see Figure 11. Elements are modeled using PShell elements with quadratic interpolation and the material is still aluminum 7075. This time, however, the compressive load is applied on one side via a rigid (RBE2) element, in the $ X $ direction. The load value is set at 120,000 lbf. The RBE2 is only active in the $ X $ direction. On the opposite side, the displacement ( $ {U}_x=0 $ ) is set to resist the compressive load. This loading boundary condition method is used as it reproduces the behavior of stiffened panels integrated in larger assemblies (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2021).
For the sizing optimization, we propose two formulations for the compressed panel. The first is the same as for the pressure bulkhead, with a compliance minimization and a weight constraint. For the second one, we use a weight minimization and a buckling load factor constraint. As for the Burst algorithm, we use the same parameters as for the pressure bulkhead case.
For the bulkhead pressure case, we have used SIMP to create an optimized baseline. However, for compressed stiffened panels, no results different from the orthogrid layout could be obtained with SIMP (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2021). Consequently, we define the baseline as the layout currently used in commercial aircraft, as illustrated in Figure 12. In Table 3, the result of a FEM-based sizing optimization is presented. The optimization considers the eigenvalues of the linear buckling as a constraint and minimizes the weight of the panel.
15. Compliance-based optimization of the compressed stiffened panels
In this run, we define the sizing optimization as a compliance minimization with a weight constraint, set to 5.0 lb. It took 22 minutes to create, size and evaluate the 500 layouts. The layouts found on the Pareto front of this run are illustrated in Figure 13, the full archive is presented as a scatter plot in Figure 14a and the convergence of compliance in Figure 14b.
The results of this run are interesting, as they reflect a particularly difficult case for topology optimization. As shown in the scatter of the results in Figure 14a, there is a plateau in the solution space. Any other solutions that do not provide stiffeners parallel to the load will both increase complexity and compliance. As such, for topology optimization methods using implicit generation (SIMP, Level-Set, MMC), it is very difficult to make any decisions that will not have a negative impact on compliance. The only topology optimization method that is capable of easily finding a solution for this case is GSM, but it is computationally very expensive due to the evaluation of buckling and the use of finite difference evaluations (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020).
CD-LEAS finds a range of efficient solutions at different complexity levels. Still, as noted earlier, buckling is often the main design driver for compressed stiffened panels. As such, to check if the results can beat the baseline, we use a different inner loop sizing to include buckling. As the panels are expected to be at least able to withstand the buckling, the linear eigenvalue constraint is set to 1.0. The objective is to reduce the weight of the panel, given the buckling load factor constraint. See Figure 15 for the results of each layout.
As these panels are sized, it appears that when properly sized for buckling, the proposed layout is unable to surpass the baseline. The best layout proposed by CD-LEAS, with only a compliance objective, has a weight of 5.5 lb., whereas the FEM-sized baseline weight 5.3 lb. This comparison is to be expected as the distribution of the stiffeners on the panel does not affect the compliance, but does significantly impact the buckling. Still, by using CD-LEAS, any unnecessary components that add complexity are ignored.
In the next section, we use an inner loop sizing optimization that considers buckling.
16. Buckling-based optimization of the compressed stiffened panels
The addition of buckling in the inner loop sizing greatly increased the computation time of CD-LEAS. On the same processor, for 500 layouts creation, sizing and evaluation took 5.8 hours. Still, a reasonable amount of time, as it can easily be run overnight, which is quite usual for structural optimization (Venkataraman & Haftka Reference Venkataraman and Haftka2004). Furthermore, compared to SIMP or GSM this computing time is a significant improvement. As discussed in (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2021), SIMP is relatively difficult to use, exploration can hardly be controlled and it requires multiple days of computation to obtain only some efficient layouts. As for GSM, a single gradient descent requires more than 5 hours of computation, with the same hardware (Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2020) for the same problem. Parallelism could reduce slightly computation time, but it would still be longer than CD-LEAS. Furthermore, exploration is only controlled by changing the initial optimization variables.
For the CD-LEAS run, the layouts found on the Pareto front are illustrated in Figure 16, the full archive is presented as a scatter in Figure 17a and the convergence of weight in Figure 17b.
Compared to the compliance-based search, using buckling shows a cleaner convergence. Also, the scatter plot shows that there are some great solutions with low complexity. Furthermore, this time the best results have a weight of 5.03, lower than the baseline with a trade-off of increased complexity. The next best solution at 5.06 still beats the baseline while having a similar complexity value to the baseline.
Choosing whether to use buckling or not during CD-LEAS search is up to the available time and the problem at hand. Still, both formulations are capable of generating feasible results with a performance similar to the ones created by experienced engineers. The compromise proposed on the Pareto front is clear and offers interesting information to the users with regards to the impact of layouts. In summary, the results obtained show a better or equal performance to the ones proposed by topology optimization, with SIMP or GSM, while requiring less computation time. More importantly, the generation used by CD-LEAS is explicit, making the representation of buckling more accurate and the exploration-focused search shows an efficient and diversified search.
17. Discussion and limitations
Both case studies have shown consistent improvement with respect to current design tools. The compliance-based solutions offered by CD-LEAS were stiff yet simple, making interpretation much easier. With our implemented complexity measure, we ensure that each stiffener behaves as independently of each other as possible (recognized Information content), that one stiffener does not dominate the performance of the panel (axiomatic Information content) and that there are no stiffeners that do not add any value (superfluous Information content). From a design point of view, our measure of complexity reflects the sensitivity of the design to parameter changes. A good physical analogy would be with a serial robotic manipulator, where you want to avoid being near a singularity (in short, a robotic singularity is a pose where joint speed becomes unexpectedly high, despite an end-effector controlled at low speed) (Angeles Reference Angeles2003). As with the robot, there exist multiple “design” singularity that needs to be avoided to reduce avoidable surprises, each in very different configurations.
For the designers, this simplicity carries into subsequent design phases where new design constraints and objectives can be integrated more easily with the new layouts. As such, the results presented for the compressed panels should not be surprising as the orthogrid layout has been proven time again that it is a very robust and efficient layout. However, this result is only obvious as the panels are usually optimized independently, separated by orthogonally placed spars and ribs. This scenario has been replicated by the boundary conditions used in this use case. As future work will increase the scale of the design space, innovative shapes are expected to arise from the use of CD-LEAS.
The case studies show that complexity measure effectively helps navigation for an effective design space exploration. By selecting a simple, albeit nonoptimal, solution as the starting point, the graph-grammar only has a few options for moves in the beginning. This smaller pool of options acts as a relaxation of the outer-loop optimization, which allows the Burst algorithm to keep looking around existing simple layouts, even if their performance is not as promising. Eventually, the process can find solutions that have both high performance and low relative complexity. Furthermore, as shown by the case studies, CD-LEAS is more effective than current topology optimization methods for the diversification of solutions, yielding multiple interesting solutions in the same run. The diversification of solutions is a significant industrial advantage, because as more design constraints arise, it can adapt and still find sound alternatives in the solution pool.
There are two advantages in the use of a relative complexity measures to limit the complexity of new layouts. First, the functional complexity is context-dependent. This effect is shown in the compressed stiffened panel case, where using only compliance still yields results that are relatively efficient in terms of buckling. Secondly, as the complexity measure is relative, there is no arbitrary threshold to set. As the process explores and builds the Pareto front at each iteration from the whole archive of solutions, it will organically set a threshold on complexity by ignoring layouts that are both complex and with low performance, as shown in both case studies.
As discussed in the introduction, the generative design framework has three main properties: generation, representation and navigation. In this work, generation is carried out using the graph-grammar, ensuring an explicit description of components and their relations through a graph. In addition, the use of the archive and the bar code ensures that the exploration does not evaluate the same layout twice. Moreover, the explicit generation is capable of creating evaluation models with an accurate representation. In this work, the representation of each layout is provided by an automatically created inner-loop optimization problem. This inner loop provides everything needed for performance and complexity assessments. As such, in terms of representation, CD-LEAS uses automation to link both topology and size optimization, which in itself is of interest for the topology optimization community. Finally, the navigation in CD-LEAS is done using the basic Burst algorithm, using both a complexity measure and any performance assessment as the objectives. In summary, by adjusting the balance of topology optimization towards exploration rather than exploitation, CD-LEAS offers an approach to layout optimization that is closer to the need of the industry. It is ready to find sound layouts and integrate new constraints that are clear compromises of simplicity and performance.
One aspect to consider, though, is that CD-LEAS trades off a reduction of complexity of the layout for an increased process complexity compared to topology optimization. SIMP topology optimization is a rather straightforward and elegant algorithm that has proven to be highly flexible when working with a simple stiffness objective. Its formulation is problem independent and physics independent. In comparison, CD-LEAS requires a more ad hoc implementation for the representation process. For the implementation discussed in this paper, we had to work with two different scripting languages and software suites that were not made specifically to work together.
In the future, there are many ways in which CD-LEAS could be improved. With regard to generation, the current implementation has only one action and few design rules. It is possible to imagine actions that add more stiffeners in a single action, such as the creation of a grid, actions that modify a grid by moving connections or still actions that combine different graphs. For example, the action “createGrid” could take as input a “spacing”, “length” and “width” parameters and create many stiffeners at once, using the same design rules as we already defined. Moreover, curvilinear stiffeners have been studied in multiple works recently, and the possibility to add them could be interesting (Mulani et al. Reference Mulani, Slemp and Kapania2013). Another possible improvement is the implementation of primary/secondary stiffeners with different heights, such as the BCF proposed in Houston et al. (Reference Houston, Quinn, Murphy and Bron2017). It could also be possible to use multiple different cross-sections. Finally, the complexity measure is based on axiomatic design, which is still debated for its practical validity (Nordlund, Lee, & Kim Reference Nordlund, Lee and Kim2015). More work could be done to improve upon the accuracy and validity of the complexity measure of CD-LEAS.
18. Conclusion
The main novelty aspect of CD-LEAS is the use of a complexity measure to drive an effective layout navigation, allowing for a more flexible explicit generation and the effective use of critical nonlinear constraints. With CD-LEAS we propose a new process for the generation of stiffened panel layouts. Leveraging knowledge acquired from generative design, axiomatic design and topology optimization, the CD-LEAS process is a two-level topology optimization process that is closer to the need of industrial structural designers and regulation aspects. This process leads to the generation of trade-offs with respect to performance and simplicity. This trade-off is important and was not evident with algorithms such as SIMP. Often times, the layouts generated required extensive design and performance analysis (Dugré et al. Reference Dugré, Vadean and Chaussée2016; Gamache et al. Reference Gamache, Vadean, Dodane and Achiche2019). As an example, in the buckling-based case study, a trade-off of less than 1% of weight can reduce complexity from 0.94 to 0.28. The results of this work are limited to a scope of smaller components as we have yet to test it on larger assemblies. Still, we finally have a process for which we can be confident, at least for small problems, that the results will allow the designers to learn something about the design space. In comparison, there are multiple SIMP projects that have simply sent the designer towards a challenging post-processing phase with no profits (Aage et al. Reference Aage, Andreassen, Lazarov and Sigmund2017). On the whole, CD-LEAS offers an improvement in both performance and process reliability and its potential for larger-scale problems is promising.
The effectiveness of CD-LEAS comes from the introduction of the complexity measure which imitates the reasoning process of an experienced designer with an open mind for the navigation of the design space. This new objective helps navigation to stay focused on simple solutions and compromises only when necessary. This approach is different from current topology optimization methods that only focus on a given objective function and will make any sacrifice necessary to complexity to get 1 or 2 final percentage of performance. As CD-LEAS considers complexity, it becomes possible to select slightly underperforming solutions, in exchange for reduced complexity which opens up the integration of possible new design constraints. Topology optimization searches only the local space of the large topological design space.
Furthermore, CD-LEAS has shown an improvement in both computation time and performance with respect to widely used topology optimization methods for the case of buckling optimization. Here again, by considering a quantifiable metric of complexity, we force the process to start with no stiffeners and thus low complexity. Increasing complexity only when necessary, also yields simpler inner loop optimization problems, consequently reducing processing time.
Improvements from CD-LEAS are presented with two case studies that are compared with previous work of the authors regarding topology optimization for stiffened panels. In the bulkhead case study, we have shown an improved exploration, a stiffer layout and multiple good trade-off solutions. For the compressed panel case study, we have shown that using CD-LEAS yields similar results from compliance and buckling-based objectives. This observation suggests a hypothesis to be validated in future work: reducing complexity yields more versatile solutions, as suggested by axiomatic design. In this particular case, previous work had identified a very difficult use with SIMP, whereas CD-LEAS has shown a consistent convergence towards an efficient minimum. Future work to explore larger and more challenging case studies will be required to assess the full potential of CD-LEAS, and more broadly the use of complexity measures in a generative design process.
Financial support
The authors wish to thank Stelia Amérique du Nord and Bombardier as industrial collaborators in this project. We would also like to thank the CARIC and MITACS for their funding of our project: MDO-1601: MUltidisciplinary Framework for Optimization of wingboX – 1a [IT07469].