CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 749

_id ga0007
id ga0007
authors Coates, Paul and Miranda, Pablo
year 2000
title Swarm modelling. The use of Swarm Intelligence to generate architectural form
source International Conference on Generative Art
summary .neither the human purposes nor the architect's method are fully known in advance. Consequently, if this interpretation of the architectural problem situation is accepted, any problem-solving technique that relies on explicit problem definition, on distinct goal orientation, on data collection, or even on non-adaptive algorithms will distort the design process and the human purposes involved.' Stanford Anderson, "Problem-Solving and Problem-Worrying". The works concentrates in the use of the computer as a perceptive device, a sort of virtual hand or "sense", capable of prompting an environment. From a set of data that conforms the environment (in this case the geometrical representation of the form of the site) this perceptive device is capable of differentiating and generating distinct patterns in its behavior, patterns that an observer has to interpret as meaningful information. As Nicholas Negroponte explains referring to the project GROPE in his Architecture Machine: 'In contrast to describing criteria and asking the machine to generate physical form, this exercise focuses on generating criteria from physical form.' 'The onlooking human or architecture machine observes what is "interesting" by observing GROPE's behavior rather than by receiving the testimony that this or that is "interesting".' The swarm as a learning device. In this case the work implements a Swarm as a perceptive device. Swarms constitute a paradigm of parallel systems: a multitude of simple individuals aggregate in colonies or groups, giving rise to collaborative behaviors. The individual sensors can't learn, but the swarm as a system can evolve in to more stable states. These states generate distinct patterns, a result of the inner mechanics of the swarm and of the particularities of the environment. The dynamics of the system allows it to learn and adapt to the environment; information is stored in the speed of the sensors (the more collisions, the slower) that acts as a memory. The speed increases in the absence of collisions and so providing the system with the ability to forget, indispensable for differentiation of information and emergence of patterns. The swarm is both a perceptive and a spatial phenomenon. For being able to Interact with an environment an observer requires some sort of embodiment. In the case of the swarm, its algorithms for moving, collision detection, and swarm mechanics conform its perceptive body. The way this body interacts with its environment in the process of learning and differentiation of spatial patterns constitutes also a spatial phenomenon. The enactive space of the Swarm. Enaction, a concept developed by Maturana and Varela for the description of perception in biological terms, is the understanding of perception as the result of the structural coupling of an environment and an observer. Enaction does not address cognition in the currently conventional sense as an internal manipulation of extrinsic 'information' or 'signals', but as the relation between environment and observer and the blurring of their identities. Thus, the space generated by the swarm is an enactive space, a space without explicit description, and an invention of the swarm-environment structural coupling. If we consider a gestalt as 'Some property -such as roundness- common to a set of sense data and appreciated by organisms or artefacts' (Gordon Pask), the swarm is also able to differentiate space 'gestalts' or spaces of some characteristics, such as 'narrowness', or 'fluidness' etc. Implicit surfaces and the wrapping algorithm. One of the many ways of describing this space is through the use of implicit surfaces. An implicit surface may be imagined as an infinitesimally thin band of some measurable quantity such as color, density, temperature, pressure, etc. Thus, an implicit surface consists of those points in three-space that satisfy some particular requirement. This allows as to wrap the regions of space where a difference of quantity has been produced, enclosing the spaces in which some particular events in the history of the Swarm have occurred. The wrapping method allows complex topologies, such as manifoldness in one continuous surface. It is possible to transform the information generated by the swarm in to a landscape that is the result of the particular reading of the site by the swarm. Working in real time. Because of the complex nature of the machine, the only possible way to evaluate the resulting behavior is in real time. For this purpose specific applications had to be developed, using OpenGL for the Windows programming environment. The package consisted on translators from DXF format to a specific format used by these applications and viceversa, the Swarm "engine", a simulated parallel environment, and the Wrapping programs, to generate the implicit surfaces. Different versions of each had been produced, in different stages of development of the work.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 39c6
authors Miranda, Pablo and Coates, Paul
year 2000
title Swarm modelling. The use of Swarm Intelligence to generate architectural form
source 3th International Conference on Generative Art, Politecnico di Milano University, Milan, Italy
summary In general the paper discusses the morphogenetic properties of swarm behaviour, and presents an example of mapping trajectories in the space of forms onto 3d flocking boids. This allows the construction of a kind of analogue to the string writing genetic algorithms and Genetic programming that are more familiar, and which have been reported by CECA. Earlier work with autonomous agents at CECA were concerned with the behaviour of agents embedded in an environment, and interactions between perceptive agents and their surrounding form. As elaborated below, the work covered in this paper is a refinement and abstraction of those experiments. This places the swarm back where perhaps it should have belonged, into the realms of abstract computation, where the emergent behaviours (the familiar flocking effect, and other observable morphologies) are used to control any number of alternative lower level morphological parameters, and to search the space of all possible variants in a directed and parallel way.
keywords Swarm Intelligence; Autonomous agents; Enactive Perception; Structural Coupling; Sensory-motor Perception; Stigmergy
series other
email
last changed 2003/03/24 17:13

_id 2ea9
authors Miranda, Pablo and Coates, Paul
year 2000
title Swarm modelling. The use of Swarm Intelligence to generate architectural form
source 4th International Conference on Generative Art, Politecnico di Milano University, Milan, Italy
summary In general the paper discusses the morphogenetic properties of swarm behaviour, and presents an example of mapping trajectories in the space of forms onto 3d flocking boids. This allows the construction of a kind of analogue to the string writing genetic algorithms and Genetic programming that are more familiar, and which have been reported by CECA. Earlier work with autonomous agents at CECA were concerned with the behaviour of agents embedded in an environment, and interactions between perceptive agents and their surrounding form. As elaborated below, the work covered in this paper is a refinement and abstraction of those experiments. This places the swarm back where perhaps it should have belonged, into the realms of abstract computation, where the emergent behaviours (the familiar flocking effect, and other observable morphologies) are used to control any number of alternative lower level morphological parameters, and to search the space of all possible variants in a directed and parallel way.
keywords Swarm Intelligence; Autonomous agents; Enactive Perception; Structural Coupling; Sensory-motor Perception; Stigmergy
series other
email
last changed 2003/03/24 15:46

_id 958e
authors Coppola, Carlo and Ceso, Alessandro
year 2000
title Computer Aided Design and Artificial Intelligence in Urban and Architectural Design
source Promise and Reality: State of the Art versus State of Practice in Computing for the Design and Planning Process [18th eCAADe Conference Proceedings / ISBN 0-9523687-6-5] Weimar (Germany) 22-24 June 2000, pp. 301-307
doi https://doi.org/10.52842/conf.ecaade.2000.301
summary In general, computer-aided design is still limited to a rather elementary use of the medium, as it is mainly used for the representation/simulation of a design idea w an electronic drawing-table. hich is not computer-generated. The procedures used to date have been basically been those of an electronic drawing-table. At the first stage of development the objective was to find a different and better means of communication, to give form to an idea so as to show its quality. The procedures used were 2D design and 3D simulation models, usually used when the design was already defined. The second stage is when solid 3D modelling is used to define the formal design at the conception stage, using virtual models instead of study models in wood, plastic, etc. At the same time in other connected fields the objective is to evaluate the feasibility of the formal idea by means of structural and technological analysis. The third stage, in my opinion, should aim to develop procedures capable of contributing to both the generation of the formal idea and the simultaneous study of technical feasibility by means of a decision-making support system aided by an Artificial Intelligence procedure which will lead to what I would describe as the definition of the design in its totality. The approach to architectural and urban design has been strongly influenced by the first two stages, though these have developed independently and with very specific objectives. It is my belief that architectural design is now increasingly the result of a structured and complex process, not a simple act of pure artistic invention. Consequently, I feel that the way forward is a procedure able to virtually represent all the features of the object designed, not only in its definitive configuration but also and more importantly in the interactions which determine the design process as it develops. Thus A.I. becomes the means of synthesis for models which are hierarchically subordinated which together determine the design object in its developmental process, supporting decision-making by applying processing criteria which generative modelling has already identified. This trend is currently being experimented, giving rise to interesting results from process design in the field of industrial production.
series eCAADe
email
more http://www.uni-weimar.de/ecaade/
last changed 2022/06/07 07:56

_id avocaad_2001_22
id avocaad_2001_22
authors Jos van Leeuwen, Joran Jessurun
year 2001
title XML for Flexibility an Extensibility of Design Information Models
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary The VR-DIS research programme aims at the development of a Virtual Reality – Design Information System. This is a design and decision support system for collaborative design that provides a VR interface for the interaction with both the geometric representation of a design and the non-geometric information concerning the design throughout the design process. The major part of the research programme focuses on early stages of design. The programme is carried out by a large number of researchers from a variety of disciplines in the domain of construction and architecture, including architectural design, building physics, structural design, construction management, etc.Management of design information is at the core of this design and decision support system. Much effort in the development of the system has been and still is dedicated to the underlying theory for information management and its implementation in an Application Programming Interface (API) that the various modules of the system use. The theory is based on a so-called Feature-based modelling approach and is described in the PhD thesis by [first author, 1999] and in [first author et al., 2000a]. This information modelling approach provides three major capabilities: (1) it allows for extensibility of conceptual schemas, which is used to enable a designer to define new typologies to model with; (2) it supports sharing of conceptual schemas, called type-libraries; and (3) it provides a high level of flexibility that offers the designer the opportunity to easily reuse design information and to model information constructs that are not foreseen in any existing typologies. The latter aspect involves the capability to expand information entities in a model with relationships and properties that are not typologically defined but applicable to a particular design situation only; this helps the designer to represent the actual design concepts more accurately.The functional design of the information modelling system is based on a three-layered framework. In the bottom layer, the actual design data is stored in so-called Feature Instances. The middle layer defines the typologies of these instances in so-called Feature Types. The top layer is called the meta-layer because it provides the class definitions for both the Types layer and the Instances layer; both Feature Types and Feature Instances are objects of the classes defined in the top layer. This top layer ensures that types can be defined on the fly and that instances can be created from these types, as well as expanded with non-typological properties and relationships while still conforming to the information structures laid out in the meta-layer.The VR-DIS system consists of a growing number of modules for different kinds of functionality in relation with the design task. These modules access the design information through the API that implements the meta-layer of the framework. This API has previously been implemented using an Object-Oriented Database (OODB), but this implementation had a number of disadvantages. The dependency of the OODB, a commercial software library, was considered the most problematic. Not only are licenses of the OODB library rather expensive, also the fact that this library is not common technology that can easily be shared among a wide range of applications, including existing applications, reduces its suitability for a system with the aforementioned specifications. In addition, the OODB approach required a relatively large effort to implement the desired functionality. It lacked adequate support to generate unique identifications for worldwide information sources that were understandable for human interpretation. This strongly limited the capabilities of the system to share conceptual schemas.The approach that is currently being implemented for the core of the VR-DIS system is based on eXtensible Markup Language (XML). Rather than implementing the meta-layer of the framework into classes of Feature Types and Feature Instances, this level of meta-definitions is provided in a document type definition (DTD). The DTD is complemented with a set of rules that are implemented into a parser API, based on the Document Object Model (DOM). The advantages of the XML approach for the modelling framework are immediate. Type-libraries distributed through Internet are now supported through the mechanisms of namespaces and XLink. The implementation of the API is no longer dependent of a particular database system. This provides much more flexibility in the implementation of the various modules of the VR-DIS system. Being based on the (supposed to become) standard of XML the implementation is much more versatile in its future usage, specifically in a distributed, Internet-based environment.These immediate advantages of the XML approach opened the door to a wide range of applications that are and will be developed on top of the VR-DIS core. Examples of these are the VR-based 3D sketching module [VR-DIS ref., 2000]; the VR-based information-modelling tool that allows the management and manipulation of information models for design in a VR environment [VR-DIS ref., 2000]; and a design-knowledge capturing module that is now under development [first author et al., 2000a and 2000b]. The latter module aims to assist the designer in the recognition and utilisation of existing and new typologies in a design situation. The replacement of the OODB implementation of the API by the XML implementation enables these modules to use distributed Feature databases through Internet, without many changes to their own code, and without the loss of the flexibility and extensibility of conceptual schemas that are implemented as part of the API. Research in the near future will result in Internet-based applications that support designers in the utilisation of distributed libraries of product-information, design-knowledge, case-bases, etc.The paper roughly follows the outline of the abstract, starting with an introduction to the VR-DIS project, its objectives, and the developed theory of the Feature-modelling framework that forms the core of it. It briefly discusses the necessity of schema evolution, flexibility and extensibility of conceptual schemas, and how these capabilities have been addressed in the framework. The major part of the paper describes how the previously mentioned aspects of the framework are implemented in the XML-based approach, providing details on the so-called meta-layer, its definition in the DTD, and the parser rules that complement it. The impact of the XML approach on the functionality of the VR-DIS modules and the system as a whole is demonstrated by a discussion of these modules and scenarios of their usage for design tasks. The paper is concluded with an overview of future work on the sharing of Internet-based design information and design knowledge.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id 53c6
authors Mardaljevic, John
year 2000
title Daylight Simulation: Validation, Sky Models and Daylight Coefficients
source De Montfort University, Leicester, UK
summary The application of lighting simulation techniques for daylight illuminance modelling in architectural spaces is described in this thesis. The prediction tool used for all the work described here is the Radiance lighting simulation system. An overview of the features and capabilities of the Radiance system is presented. Daylight simulation using the Radiance system is described in some detail. The relation between physical quantities and the lighting simulation parameters is made clear in a series of progressively more complex examples. Effective use of the inter-reflection calculation is described. The illuminance calculation is validated under real sky conditions for a full-size office space. The simulation model used sky luminance patterns that were based directly on measurements. Internal illuminance predictions are compared with measurements for 754 skies that cover a wide range of naturally occurring conditions. The processing of the sky luminance measurements for the lighting simulation is described. The accuracy of the illuminance predictions is shown to be, in the main, comparable with the accuracy of the model input data. There were a number of predictions with low accuracy. Evidence is presented to show that these result from imprecision in the model specification - such as, uncertainty of the circumsolar luminance - rather than the prediction algorithms themselves. Procedures to visualise and reduce illuminance and lighting-related data are presented. The ability of sky models to reproduce measured sky luminance patterns for the purpose of predicting internal illuminance is investigated. Four sky models and two sky models blends are assessed. Predictions of internal illuminance using sky models/blends are compared against those using measured sky luminance patterns. The sky model blends and the Perez All-weather model are shown to perform comparably well. Illuminance predictions using measured skies however were invariably better than those using sky models/blends. Several formulations of the daylight coefficient approach for predicting time varying illuminances are presented. Radiance is used to predict the daylight coefficients from which internal illuminances are derived. The form and magnitude of the daylight coefficients are related to the scene geometry and the discretisation scheme. Internal illuminances are derived for four daylight coefficient formulations based on the measured luminance patterns for the 754 skies. For the best of the formulations, the accuracy of the daylight coefficient derived illuminances is shown to be comparable to that using the standard Radiance calculation method. The use of the daylight coefficient approach to both accurately and efficiently predict hourly internal daylight illuminance levels for an entire year is described. Daylight coefficients are invariant to building orientation for a fixed building configuration. This property of daylight coefficients is exploited to yield hourly internal illuminances for a full year as a function of building orientation. Visual data analysis techniques are used to display and process the massive number of derived illuminances.
series thesis:PhD
email
more http://www.iesd.dmu.ac.uk/~jm/thesis/
last changed 2003/02/12 22:37

_id 44be
authors Radford, Antony
year 2000
title Games and learning about form in architecture
source Automation in Construction 9 (4) (2000) pp. 379-385
summary Metaphors of play and games in the use of digital media in architectural education are described. The essence of 'games' lies in a person's immersion in play subject to rules. The paper argues that embracing these metaphors is an effective means of developing student confidence and abilities in spatial modelling, design composition, and form creation. Some examples are provided.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:23

_id 83cb
authors Telea, Alexandru C.
year 2000
title Visualisation and simulation with object-oriented networks
source Eindhoven University of Technology
summary Among the existing systems, visual programming environments address best these issues. However, producing interactive simulations and visualisations is still a difficult task. This defines the main research objective of this thesis: The development and implementation of concepts and techniques to combine visualisation, simulation, and application construction in an interactive, easy to use, generic environment. The aim is to produce an environment in which the above mentioned activities can be learnt and carried out easily by a researcher. Working with such an environment should decrease the amount of time usually spent in redesigning existing software elements such as graphics interfaces, existing computational modules, and general infrastructure code. Writing new computational components or importing existing ones should be simple and automatic enough to make using the envisaged system an attractive option for a non programmer expert. Besides this, all proven successful elements of an interactive simulation and visualisation environment should be provided, such as visual programming, graphics user interfaces, direct manipulation, and so on. Finally, a large palette of existing scientific computation, data processing, and visualisation components should be integrated in the proposed system. On one hand, this should prove our claims of openness and easy code integration. On the other hand, this should provide the concrete set of tools needed for building a range of scientific applications and visualisations. This thesis is structured as follows. Chapter 2 defines the context of our work. The scientific research environment is presented and partitioned into the three roles of end user, application designer, and component developer. The interactions between these roles and their specific requirements are described and lead to a more precise formulation of our problem statement. Chapter 3 presents the most used architectures for simulation and visualisation systems: the monolithic system, the application library, and the framework. The advantages and disadvantages of these architectural models are then discussed in relation with our problem statement requirements. The main conclusion drawn is that no single existing architectural model suffices, and that what is needed is a combination of the features present in all three models. Chapter 4 introduces the new architectural model we propose, based on the combination of object-orientation in form of the C++ language and dataflow modelling in the new MC++ language. Chapter 5 presents VISSION, an interactive simulation and visualisation environment constructed on the introduced new architectural model, and shows how the usual tasks of application construction, steering, and visualisation are addressed. In chapter 6, the implementation of VISSION’s architectural model is described in terms of its component parts. Chapter 7 presents the applications of VISSION to numerical simulation, while chapter 8 focuses on its visualisation and graphics applications. Finally, chapter 9 concludes the thesis and outlines possible direction for future research.
keywords Computer Visualisation
series thesis:PhD
email
last changed 2003/02/12 22:37

_id 60a7
authors Monedero, Javier
year 2000
title Parametric design: a review and some experiences
source Automation in Construction 9 (4) (2000) pp. 369-377
summary During the last few years there has been an extraordinary development of computer-aided tools intended to present or communicate the results of architectural projects. But there has not been a comparable progress in the development of tools intended to assist design to generate architectural forms in an easy and interactive way. Even worse, architects who use the powerful means provided by computers as a direct tool to create architectural forms are still an exception. Architecture continues to be produced by traditional means using the computer as little more than a drafting tool. The main reasons that may explain this situation can be identified rather easily, although there will be significant differences of opinion. In my opinion, it is a mistake trying to advance too rapidly and, for instance, proposing integrated design methods using expert systems and artificial intelligence while no adequate tools to generate and modify simple 3D-models are available. The modeling tools we have at the present moment are unsatisfactory. Their principal limitation is the lack of appropriate instruments to modify interactively the model once it has been created. This is a fundamental aspect in any design activity, where the designer is constantly going forward and backwards, re-elaborating once and again some particular aspect of the model, or its general layout, or even coming back to a previous solution that had been temporarily abandoned. This paper presents a general summary of the actual situation and recent developments that may be incorporated to architectural design tools in a near future, together with some critical remarks about their relevance to architecture.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:22

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id a136
authors Blaise, J.Y., Dudek, I. and Drap, P.
year 1998
title Java collaborative interface for architectural simulations A case study on wooden ceilings of Krakow
source International Conference On Conservation - Krakow 2000, 23-24 November 1998, Krakow, Poland
summary Concern for the architectural and urban preservation problems has been considerably increasing in the past decades, and with it the necessity to investigate the consequences and opportunities opened for the conservation discipline by the development of computer-based systems. Architectural interventions on historical edifices or in preserved urban fabric face conservationists and architects with specific problems related to the handling and exchange of a variety of historical documents and representations. The recent development of information technologies offers opportunities to favour a better access to such data, as well as means to represent architectural hypothesis or design. Developing applications for the Internet also introduces a greater capacity to exchange experiences or ideas and to invest on low-cost collaborative working platforms. In the field of the architectural heritage, our research addresses two problems: historical data and documentation of the edifice, methods of representation (knowledge modelling and visualisation) of the edifice. This research is connected with the ARKIW POLONIUM co-operation program that links the MAP-GAMSAU CNRS laboratory (Marseilles, France) and the Institute HAiKZ of Kraków's Faculty of Architecture. The ARKIW programme deals with questions related to the use of information technologies in the recording, protection and studying of the architectural heritage. Case studies are chosen in order to experience and validate a technical platform dedicated to the formalisation and exchange of knowledge related to the architectural heritage (architectural data management, representation and simulation tools, survey methods, ...). A special focus is put on the evolution of the urban fabric and on the simulation of reconstructional hypothesis. Our contribution will introduce current ARKIW internet applications and experiences: The ARPENTEUR architectural survey experiment on Wieża Ratuszowa (a photogrammetrical survey based on an architectural model). A Gothic and Renaissance reconstruction of the Ratusz Krakowski using a commercial modelisation and animation software (MAYA). The SOL on line documentation interface for Kraków's Rynek G_ówny. Internet analytical approach in the presentation of morphological informations about Kraków's Kramy Bogate Rynku Krakowskiego. Object-Orientation approach in the modelling of the architectural corpus. The VALIDEUR and HUBLOT Virtual Reality modellers for the simulation and representation of reconstructional hypothesis and corpus analysis.
series other
last changed 2003/04/23 15:14

_id cb42
authors Coates, Paul and Schmid, Claudia
year 1999
title Agent Based Modelling
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 652-661
doi https://doi.org/10.52842/conf.ecaade.1999.652
summary This paper describes the work of students in Unit 6 of the Diploma school at the UEL during 1998-9. The unit in association with the MSc has been exploring ways of using the computer to explore the idea of emergent form as a way of generating designs, and a way of focussing the pedagogic process on a new and interesting set of determinants of form.
keywords Agent Based Modelling, Generative Modelling, Emergent Forms
series eCAADe
email
last changed 2022/06/07 07:56

_id de43
authors Counsell, J.
year 2000
title The management and visualisation of 3-dimensional models using a spatial database
source CIDAC, Volume 2 Issue 4 November 2000, pp. 225-235
summary Each year, computer-aided architectural design (CAAD) systems in common use are enhanced and gain facilities that ease 3-dimensional (3D) modelling. Consequently, large complex datasets are increasinly common during the creation and management of 3D models of buildings and urban areas. Uses for such models range from the automatic generation of drawings and schedules to virtual reality (VR) and visualisation across the web. Geographic information systems (GISs) are optimised for the management and retrieval of spatial data and may be used to assist both management and visualisation of large 3D datasets using open standards, such as the ISO standard virtual reality modelling langauge (VRML). Experience gained in the use of such systems indicates a need for specific procedures for recording 3D data and creating linkages to other information. It is suggested that these procedures are applicable to a broad range of such models.
keywords VRML, Urban 3-Dimensional Models, GIS Management
series journal paper
last changed 2003/05/15 21:23

_id 9b63
authors De Mesa, A., Quilez, J. and Regot, J.
year 1999
title Sunlight Energy Graphic and Analytic Control in 3D Modelling
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 733-738
doi https://doi.org/10.52842/conf.ecaade.1999.733
summary Linking solar positions with architecture is a traditional idea, but the use of graphical tools to control sunlight in urban surroundings or buildings is relatively recent. A three-dimensional working environment like the computer offers a new dimension to verify the relationships between the sun and the architecture. This paper shows a new way to calculate the incidence of solar energy in architectural environments using computer 3D modelling. The addition of virtual space visualisation to the analytic computation brings a new tool that simplifies the technical study of sunlight. We have developed several programs based upon the three-dimensional construction of the solar vault and the obstructing objects for a defined position. The first one draws the solar vault for a defined range of dates according to latitude, that is the basis of the energetic calculation. The second program computes the obstruction, i.e. the solar regions that are obstructed by any object. Finally, the third one, allow us to define an orientation to compute the energy that arrives to the analysed positioning. The last program returns the result of calculation in several ways: it shows the amount of energy through colours and makes a list of solar hours according to its energy.
keywords Sunlight, Energy, 3D modelling
series eCAADe
last changed 2022/06/07 07:56

_id b1db
authors Francis, Sabu
year 1999
title The Importance of Being Abstract: An Indian Approach to Models
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 101-109
doi https://doi.org/10.52842/conf.ecaade.1999.101
summary Traditional Indian way of life is surrounded by ambiguity. This is in direct contrast to an Aristotelian approach, where polarised stands are always taken. A black and white approach tends to yield results speedily, but exhaustive solutions which can explain complexity are usually brute force procedures. Even so, their conclusions in the end are still suspect. The author believes that rich solutions may exist when we use an 'alternate' or abstract synthesized reality to do our modelling instead of relying on analogies and other direct links to the real world. Models that allow synthesis tend to accept ambiguity. The author presents in this paper an 'unconventional' system to represent architecture which has had some amount of success probably because it started of, on pure abstract grounds that allowed ambiguity instead of basing it on an Aristotelian, analytical model.
keywords Aristotle, Buddha, Representations, Abstract Models
series eCAADe
email
last changed 2022/06/07 07:50

_id ddssar0012
id ddssar0012
authors Hendricx, Ann and Neuckermans, Herman
year 2000
title Setting objects to work: adding functionality to an architectural object model
source Timmermans, Harry (Ed.), Fifth Design and Decision Support Systems in Architecture and Urban Planning - Part one: Architecture Proceedings (Nijkerk, the Netherlands)
summary Several research initiatives in the field of product modelling have produced static descriptions of the architectural and geometrical objects capable of describing architectural design projects. Less attention is paid to the development phase in which these static models are transformed into workable architectural design environments. In the context of the IDEA+ research project (Integrated Design Environment for Architectural Design), we use the object-oriented analysis method MERODE to develop and describe both an enterprise (or product) model and a functionality model. On the one hand, the enterprise model defines the architectural and geometrical objects, their methods and their relation with other objects. On the other hand, the functionality model organizes the functionality objects – ranging from single-event objects to complex-workflow objects – in a layered and easily expandable system. The functionality model is created on top of the enterprise model and closes the gap between the static enterprise model and the dynamic design environment as a whole. After a short introduction of the envisaged design environment and its underlying enterprise model, the paper will concentrate on the presentation of the higher-level functionality model. Elaborated examples of functionality objects on the different levels will clarify its concepts and proof its feasibility.
series DDSS
last changed 2003/08/07 16:36

_id 5266
authors Jakimowicz, Adam
year 1999
title An Intuitive CAAD
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 80-85
doi https://doi.org/10.52842/conf.ecaade.1999.080
summary This paper presents the educational experiment, which emerged from the junction of different inspirations and needs. It joins the experiences from an individual research on intuitive 3d computer modelling, courses of traditional architectural composition and the idea of individualisation of the computer use in architectural design education. It shows how computers were used in part of architectural education in a non-computer-oriented course. The experiment was included to and further developed within the frame of AVOCAAD project.
series eCAADe
email
last changed 2022/06/07 07:52

_id 55ca
authors Mase, Jitsuro
year 2000
title Moderato: 3D Sketch CAD with Quick Positioned Working Plane and Texture Modelling
source Promise and Reality: State of the Art versus State of Practice in Computing for the Design and Planning Process [18th eCAADe Conference Proceedings / ISBN 0-9523687-6-5] Weimar (Germany) 22-24 June 2000, pp. 269-272
doi https://doi.org/10.52842/conf.ecaade.2000.269
summary The lack of computer systems that can be easily used during the early stages of the architectural design process has been discussed for many years. The usual argument starts with the recognition that hand drawn sketches are an important tool in the early stage of both professional and student design because they can be used to visualise the designer’s ideas quickly and have the flexibility to handle any shape the designer imagines. Research has then mostly focused on using computer based sketch recognition to directly produce three dimensional models from hand drawn sketches. However sketch recognition still has certain problems that require the drawing action of users to be constrained in some way in order to be solved. If sketch recognition is still imperfect, the possibility of directly sketching within digital 3D space should be considered. Some systems allowing user to sketch in digital 3D space have been developed which do not depend on sketch recognition. Although Piranesi does not aim to support sketch design, it does allow the user to paint in the Z-buffer space - an unique idea termed "interactive rendering." SketchVRML tries to generate 3D geometrical data automatically from 2D hand drawn sketches by adding the depth value to the drawn lines according to the strength of line strokes. SketchBoX provides translucent surfaces in digital 3D space which can be glued onto existing objects or arranged anywhere in space. These surfaces have texture map data which can be modified by painting onto the texture. Transparent textures can be painted onto the surfaces to create see-through portions. Moderato also uses this technique to model a polygonŐs shape.
keywords Sketch, Early Stage, Interface, 3D Modelling
series eCAADe
email
more http://www.uni-weimar.de/ecaade/
last changed 2022/06/07 07:59

_id 3582
authors Maver, Th.W., Ennis, G. and Jarvis, G.
year 2001
title CHRONICLE OF THE CITY
source SIGraDi biobio2001 - [Proceedings of the 5th Iberoamerican Congress of Digital Graphics / ISBN 956-7813-12-4] Concepcion (Chile) 21-23 november 2001, pp. 322-324
summary The great cities of all continents have a remarkable story to tell about their social, economic, industrial, cultural and architectural evolution. This paper gives an insight into an ambitious project to construct a 2000 year historical chronicle of Glasgow – the city at the heart of the Industrial Revolution in Northern Europe. The multimedia document which holds the chronicle draws from a huge range of paperbased resources and pulls together contributions from experts in many disparate fields. A significant characteristic of the chronicle is the use of “place” in the telling of the story. This has been achieved by sophisticated merging of artistic impressions, digital terrain modelling and aerial photography.
series SIGRADI
email
last changed 2016/03/10 09:55

_id ecaade2023_205
id ecaade2023_205
authors Meeran, Ahmed and Joyce, Sam
year 2023
title Rethinking Airport Spatial Analysis and Design: A GAN based data driven approach using latent space exploration on aerial imagery for adaptive airport planning
source Dokonal, W, Hirschberg, U and Wurzer, G (eds.), Digital Design Reconsidered - Proceedings of the 41st Conference on Education and Research in Computer Aided Architectural Design in Europe (eCAADe 2023) - Volume 2, Graz, 20-22 September 2023, pp. 501–510
doi https://doi.org/10.52842/conf.ecaade.2023.2.501
summary Airports require long term planning, balancing estimations of future demand against available airfield land and site constraints. This is becoming more critical with climate change and the transition to sustainable aviation fuelling infrastructure. This paper demonstrates a novel procedure using Satellite Imagery and Generative Learning to aid in the comparative analysis and early-stage airfield design. Our workflow uses a GAN trained on 2000 images of airports transforming them into a high-dimensional latent space capturing the typologies’ large-scale features. Using a process of projection and dimensional-reduction methods we can locate real-world airport images in the generative latent space and vice-versa. With this capability we can perform comparative “neighbour” analysis at scale based on spatial similarity of features like airfield configuration, and surrounding context. Using this low-dimensional 3D ‘airport designs space’ with meaningful markers provided by existing airports allows for ‘what if’ modelling, such as visualizing an airport on a site without one, modifying an existing airport towards another target airport, or exploring changes in terrain, such as due to climate change or urban development. We present this method a new way to undertake case study, site identification and analysis, as well as undertake speculative design powered by typology informed ML generation, which can be applied to any typologies which could use aerial images to categorize them.
keywords Airport Development, Machine Learning, GAN, High Dimensional Analysis, Parametric Space Exploration, tSNE, Latent Space Exploration, Data Driven Planning
series eCAADe
email
last changed 2023/12/10 10:49

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 37HOMELOGIN (you are user _anon_266296 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002