CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 613

_id ga9916
id ga9916
authors Elzenga, R. Neal and Pontecorvo, Michael S.
year 1999
title Arties: Meta-Design as Evolving Colonies of Artistic Agents
source International Conference on Generative Art
summary Meta-design, the act of designing a system or species of design instead of a design instance, is an important concept in modern design practice and in the generative design paradigm. For meta-design to be a useful tool, the designer must have more formal support for both design species definition/expression and the abstract attributes which the designer is attempting to embody within a design. Arties is an exploration of one possible avenue for supporting meta-design. Arties is an artistic system emphasizing the co-evolution of colonies of Artificial Life design or artistic agents (called arties) and the environment they inhabit. Generative design systems have concentrated on biological genetics metaphors where a population of design instances are evolved directly from a set of ‘parent’ designs in a succession of generations. In Arties, the a-life agent which is evolved, produces the design instance as a byproduct of interacting with its environment. Arties utilize an attraction potential curve as their primary dynamic. They sense the relative attraction of entities in their environment, using multiple sensory channels. Arties then associate an attractiveness score to each entity. This attractiveness score is combined with a 'taste' function built into the artie that is sensitized to that observation channel, entity, and distance by a transfer function. Arties use this attraction to guide decisions and behaviors. A community of arties, with independent evolving attraction criteria can pass collective judgement on each point in an art space. As the Artie moves within this space it modifies the environment in reaction to what it senses. Arties support for Meta-design is in (A) the process of evolving arties, breeding their attraction potential curve parameters using a genetic algorithm and (B) their use of sensory channels to support abstract attributes geometries. Adjustment of these parameters tunes the attraction of the artie along various sensing channels. The multi-agent co-evolution of Arties is one approach to creating a system for supporting meta-design. Arties is part of an on-going exploration of how to support meta-design in computer augmented design systems. Our future work with Arties-like systems will be concerned with applications in areas such as modeling adaptive directives in Architecture, Object Structure Design, spatio-temporal behaviors design (for games and simulations), virtual ambient spaces, and representation and computation of abstract design attributes.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id b4d2
authors Caldas, Luisa G. and Norford, Leslie K.
year 1999
title A Genetic Algorithm Tool for Design Optimization
source Media and Design Process [ACADIA ‘99 / ISBN 1-880250-08-X] Salt Lake City 29-31 October 1999, pp. 260-271
doi https://doi.org/10.52842/conf.acadia.1999.260
summary Much interest has been recently devoted to generative processes in design. Advances in computational tools for design applications, coupled with techniques from the field of artificial intelligence, have lead to new possibilities in the way computers can inform and actively interact with the design process. In this paper we use the concepts of generative and goal-oriented design to propose a computer tool that can help the designer to generate and evaluate certain aspects of a solution towards an optimized behavior of the final configuration. This work focuses mostly on those aspects related to the environmental performance of the building. Genetic Algorithms are applied as a generative and search procedure to look for optimized design solutions in terms of thermal and lighting performance in a building. The Genetic Algorithm (GA) is first used to generate possible design solutions, which are then evaluated in terms of lighting and thermal behavior using a detailed thermal analysis program (DOE2.1E). The results from the simulations are subsequently used to further guide the GA search towards finding low-energy solutions to the problem under study. Solutions can be visualized using an AutoLisp routine. The specific problem addressed in this study is the placing and sizing of windows in an office building. The same method is applicable to a wide range of design problems like the choice of construction materials, design of shading elements, or sizing of lighting and mechanical systems for buildings.
series ACADIA
email
last changed 2022/06/07 07:54

_id 3ddc
authors Dijkstra, Jan and Timmermans, Harry
year 1999
title Towards a Multi-Agent Model for Visualizing Simulated User Behavior to Support the Assessment of Design Performance
source Media and Design Process [ACADIA ‘99 / ISBN 1-880250-08-X] Salt Lake City 29-31 October 1999, pp. 226-237
doi https://doi.org/10.52842/conf.acadia.1999.226
summary We introduce the outline of a multi-agent model that can be used for visualizing simulated user behavior to support the assessment of design performance. We will consider various performance indicators of building environments, which are related to user reaction to design decisions. This system may serve as a media tool in the design process for a better understanding of what the design will look like, especially for those cases where design or planning decisions will affect the behavior of individuals. The system is based on cellular automata and multi-agent simulation technology. The system simulates how agents move around in a particular 3D (or 2D) environment, in which space is represented as a lattice of cells. Agents represent objects or people with their own behavior, moving over the network. Each agent will be located in a simulated space, based on the cellular automata grid. Each iteration of the simulation is based on a parallel update of the agents conforming local rules. Agents positioned within an environment will need sensors to perceive their local neighborhood and some means with which to affect the environment. In this way, autonomous individuals and the interaction between them can be simulated by the system. As a result, designers can use the system to assess the likely consequences of their design decisions on user behavior. We think that the system provides a potentially valuable tool to support design and decision-making processes, related to user behavior in architecture and urban planning.
series ACADIA
email
last changed 2022/06/07 07:55

_id 221d
authors Lee, Sanghyun
year 1999
title Internet-based collaborative design evaluation : an architect's perspective
source Harvard University
summary This research aims at developing a design evaluation system that employs a Product Model as the logical basis for integrating building design and construction processes. The system is implemented with Java language, which allows the system to work over the Internet. Accordingly, the system helps architects to collaborate with remote participants. Thus, this design evaluation system is a building performance evaluator like DOE-2, RADIANCE, HVAC, and the Automated Building Code Checker. This research, however, is mainly concerned with an architect's view during the schematic design and design development stage, while the existing design evaluation systems cover other special consultants' views such as those of HVAC designers, structural engineers, and contractors. From an architect's view, this evaluation system checks the compliance of design objects represented by means of physical objects such as walls and windows and conceptual objects such as rooms as well, to the design criteria focused on accommodating human behavior, rather than other building performances such as sustaining building structures and maintaining indoor livability. As such, the system helps designers analyze and evaluate design solutions according to their original intent. The innovative points of this research lie in the following: (1) Unlike other inquiries, it addresses a systematic evaluation of building design from an architect's view focusing on the experiential quality of the built environment. This research demonstrates that such an evaluation becomes available by introducing human activity-based evaluation. (2) It can take a multi-agenda for several groups of different interests by providing an Aspect Model based on human activity-centered systematic translation of their design considerations and 3D model-based graphical representations into system readable ones. (3) As a result, it addresses the possibility of expanding the capability of the design evaluator from a mere code checker to a general design evaluator while simultaneously, enhancing the availability from stand alone to Internet based networking.
keywords Architectural Rendering; Data Processing; Evaluation; Buildings; Performance; World Wide Web; Internet
series thesis:PhD
last changed 2003/02/12 22:37

_id 6810
authors Makkonen, Petri
year 1999
title On multi body systems simulation in product design
source KTH Stockholm
summary The aim of this thesis is to provide a basis for efficient modelling and software use in simulation driven product development. The capabilities of modern commercial computer software for design are analysed experimentally and qualitatively. An integrated simulation model for design of mechanical systems, based on four different "simulation views" is proposed: An integrated CAE (Computer Aided Engineering) model using Solid Geometry (CAD), Finite Element Modelling (FEM), Multi Body Systems Modelling (MBS) and Dynamic System Simulation utilising Block System Modelling tools is presented. A theoretical design process model for simulation driven design based on the theory of product chromosome is introduced. This thesis comprises a summary and six papers. Paper A presents the general framework and a distributed model for simulation based on CAD, FEM, MBS and Block Systems modelling. Paper B outlines a framework to integrate all these models into MBS simulation for performance prediction and optimisation of mechanical systems, using a modular approach. This methodology has been applied to design of industrial robots of parallel robot type. During the development process, from concept design to detail design, models have been refined from kinematic to dynamic and to elastodynamic models, finally including joint backlash. A method for analysing the kinematic Jacobian by using MBS simulation is presented. Motor torque requirements are studied by varying major robot geometry parameters, in dimensionless form for generality. The robot TCP (Tool Center Point) path in time space, predicted from elastodynamic model simulations, has been transformed to the frequency space by Fourier analysis. By comparison of this result with linear (modal) eigen frequency analysis from the elastodynamic MBS model, internal model validation is obtained. Paper C presents a study of joint backlash. An impact model for joint clearance, utilised in paper B, has been developed and compared to a simplified spring-damper model. The impact model was found to predict contact loss over a wider range of rotational speed than the spring-damper model. Increased joint bearing stiffness was found to widen the speed region of chaotic behaviour, due to loss of contact, while increased damping will reduce the chaotic range. The impact model was found to have stable under- and overcritical speed ranges, around the loss of contact region. The undercritical limit depends on the gravitational load on the clearance joint. Papers D and E give examples of the distributed simulation model approach proposed in paper A. Paper D presents simulation and optimisation of linear servo drives for a 3-axis gantry robot, using block systems modelling. The specified kinematic behaviour is simulated with multi body modelling, while drive systems and control system are modelled using a block system model for each drive. The block system model has been used for optimisation of the transmission and motor selection. Paper E presents an approach for re-using CAD geometry for multi body modelling of a rock drilling rig boom. Paper F presents synthesis methods for mechanical systems. Joint and part number synthesis is performed using the Grübler and Euler equations. The synthesis is continued by applying the theory of generative grammar, from which the grammatical rules of planar mechanisms have been formulated. An example of topological synthesis of mechanisms utilising this grammar is presented. Finally, dimensional synthesis of the mechanism is carried out by utilising non-linear programming with addition of a penalty function to avoid singularities.
keywords Simulation; Optimisation; Control Systems; Computer Aided Engineering; Multi Body Systems; Finite Element Method; Backslash; Clearance; Industrial Robots; Parallel Robots
series thesis:PhD
last changed 2003/02/12 22:37

_id acadia10_183
id acadia10_183
authors Ireland, Tim
year 2010
title Stigmergic Planning
source ACADIA 10: LIFE in:formation, On Responsive Information and Variations in Architecture [Proceedings of the 30th Annual Conference of the Association for Computer Aided Design in Architecture (ACADIA) ISBN 978-1-4507-3471-4] New York 21-24 October, 2010), pp. 183-189
doi https://doi.org/10.52842/conf.acadia.2010.183
summary This paper presents an application of swarm intelligence towards the problem of spatial configuration. The methodology classifies activities as discrete entities, which self-organise topologically through associational parameters: an investigation of emergent route formation and spatial connectivity based on simple agent and pheromone interaction, coupled with the problem of ‘loose’ rectangular geometric assembly. A concept model sniffingSpace (Ireland, 2009) developed in Netlogo (Willensky, 1999), which established the self-organising topological capacity of the system, is extended in Processing (Fry & Rea, 2009) to incorporate rectangular geometry towards the problem of planning architectural space.
series ACADIA
type normal paper
email
last changed 2022/06/07 07:50

_id ga9922
id ga9922
authors Annunziato, M. and Pierucci, P.
year 1999
title The Art of Emergence
source International Conference on Generative Art
summary Since several years, the term emergence is mentioned in the paradigm of chaos and complexity. Following this approach, complex system constituted by multitude of individual develop global behavioral properties on the base of local chaotic interactions (self-organization). These theories, developed in scientific and philosophical milieus are rapidly spreading as a "way of thinking" in the several fields of cognitive activities. According to this "way of thinking" it is possible revise some fundamental themes as the economic systems, the cultural systems, the scientific paths, the communication nets under a new approach where nothing is pre-determined, but the global evolution is determined by specific mechanisms of interaction and fundamental events (bifurcation). With a jump in scale of the life, also other basic concepts related to the individuals as intelligence, consciousness, psyche can be revised as self-organizing phenomena. Such a conceptual fertility has been the base for the revision of the artistic activities as flexible instruments for the investigation of imaginary worlds, metaphor of related real worlds. In this sense we claim to the artist a role of "researcher". Through the free exploration of new concepts, he can evoke qualities, configurations and hypothesis which have an esthetical and expressive value and in the most significant cases, they can induce nucleation of cultural and scientific bifurcation. Our vision of the art-science relation is of cooperative type instead of the conflict of the past decades. In this paper we describe some of the most significant realized artworks in order to make explicit the concepts and basic themes. One of the fundamental topics is the way to generate and think to the artwork. Our characterization is to see the artwork not as a static finished product, but as an instance or a dynamic sequence of instances of a creative process which continuously evolves. In this sense, the attention is focused on the "generative idea" which constitutes the envelop of the artworks generable by the process. In this approach the role of technology (computers, synthesizers) is fundamental to create the dimension of the generative environment. Another characterizing aspect of our artworks is derived by the previous approach and specifically related to the interactive installations. The classical relation between artist, artwork and observers is viewed as an unidirectional flux of messages from the artist to the observer through the artwork. In our approach artist, artwork and observer are autonomous entities provided with own personality which jointly intervene to determine the creative paths. The artist which generate the environment in not longer the "owner" of the artwork; simply he dialectically bring the generative environment (provided by a certain degree of autonomy) towards cultural and creative "void" spaces (not still discovered). The observers start from these platforms to generate other creative paths, sometimes absolutely unexpected , developing their new dialectical relations with the artwork itself. The results derived by these positions characterize the expressive elements of the artworks (images, sequences and sounds) as the outcomes of emergent behavior or dynamics both in the sense of esthetical shapes emergent from fertile generative environments, either in terms of emergent relations between artist, artwork and observer, either in terms of concepts which emerge by the metaphor of artificial worlds to produce imaginary hypothesis for the real worlds.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id e15c
authors Bartenbach, Christian and Witting, Walter
year 1999
title VDU WORK IN DIFFERENT LIGHTING CONDITIONS
source Full-scale Modeling and the Simulation of Light [Proceedings of the 7th European Full-scale Modeling Association Conference / ISBN 3-85437-167-5] Florence (Italy) 18-20 February 1999, pp. 7-28
summary In order to avoid the disadvantages of purely subjective methods in a technical evaluation of daylight and artificial light systems, the Bartenbach LichtLabor developed new test methods which can determine objectively and quantitatively the visual or psycho-physiological stress connected with VDU work [1], depending on different lighting conditions. Daylight and artificial lighting systems were tested with these methods and compared by using the performances achieved by the test subjects. Some highly significant differences in performance done under the individual lighting systems became apparent and demonstrated that the visual stress or the physical or physiological fatigue from an ergonomic viewpoint depends largely on the lighting conditions at the workplace. This holds true for daylight systems (glare protection, re-directing lamellae, clear window as a control condition) as well as for purely artificial lighting systems where especially the choice of color temperature of the light and the used control gear (conventional or electronic) determine the resulting performance. Optimized lighting also positively affects the productivity and economicy for the design of workplaces that take the human factor into account.
keywords VDU, Optimized Lighting, Performance Test, Lighting System, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 11:27

_id 3017
authors Carson, J. and Clark, A.
year 1999
title Multicast Shared Virtual Worlds Using VRML 97
source Proceedings of VRML 99 Fourth Symposium on the Virtual Reality Modeling language, The Association for Computing Machinery, Inc. New York, pp. 133-140
summary This paper describes a system for authoring and executing shared virtual worlds within existing VRML97 viewers such as Cosmo Player. As VRML97 does not contain any direct support for the construction of virtual worlds containing multiple users extensions are presented to provide support for shared behaviours, avatars and objects that can be manipulated and carried by participants in the world; these extensions are pre-processed into standard VRML97 and Java. A system infrastructure is described which allows worlds to be authored and executed within the context of the World Wide Web and the MBone. CR Categories and Subject Descriptors: C.2.2 [Computer Communication Networks]: Network Protocols - Applications; C.2.4 [Computer Communication Networks]: Distributed Sys- tems - Distributed Applications; H.5.1 [Information Interfaces and Presentation] Multimedia Information Systems - Artificial, Aug- mented and Virtual Realities; 1.3.2 [Computer Graphics]: Graphics Systems - Distributed/network graphics: 1.3.6 [Computer Graph- ics]: Methodology and Techniques - Interaction Techniques; 1.3.7 [Computer Graphics]: Three Dimensional Graphics and Realism - Virtual Reality.
series other
last changed 2003/04/23 15:50

_id 37d1
authors Corona Martíne, Alfonso and Vigo, Libertad
year 1999
title Before the Digital Design Studio
source III Congreso Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings] Montevideo (Uruguay) September 29th - October 1st 1999, pp. 247-252
summary This paper contains some observations which derive from our work as Studio Professors . In the last years, studios are in a transition phase with the progressive introduction of computers in later stages of the design process. The initiative generally belongs to students rather than to studio masters, since the former are aware that a knowledge of CAD systems will make them able to get work in architects offices. It is the first few Studios that will guide the student in forming a conception of what is architecture . Therefore, we have observer more attentively the way in which he establishes his first competence as a designer. We believe it is useful to clarify design training before we can integrate computers into it. The ways we all learn to design and which we transmit in the Studio were obviously created a long time ago, when Architecture became a subject taught in Schools, no longer a craft to be acquired under a master. The conception of architecture that the student forms in his mind is largely dependent on a long tradition of Beaux-Arts training which survives (under different forms) in Modern Architecture. The methods he or she acquires will become the basis of his creative design process also in professional life. Computer programmes are designed to adapt into the stages of this design process simply as time saving tools. We are interested in finding out how they can become an active part in the creative process and how to control this integration in teaching. Therefore, our work deals mainly with the tradition of the Studio and the conditioning it produces. The next step will be to explore the possiblities and restrictions that will inevitably issue from the introduction of new media.
series SIGRADI
email
last changed 2016/03/10 09:49

_id c408
authors Hirschberg, U., Schmitt, G., Kurmann, D., Kolarevic, B., Johnson, B. and Donath, D.
year 1999
title The 24 Hour Design Cycle: An Experiment in Design Collaboration over the Internet
source CAADRIA '99 [Proceedings of The Fourth Conference on Computer Aided Architectural Design Research in Asia / ISBN 7-5439-1233-3] Shanghai (China) 5-7 May 1999, pp. 181-190
doi https://doi.org/10.52842/conf.caadria.1999.181
summary This paper describes a Virtual Design Studio exercise involving three academic institutions University - Hong Kong (China), ETH Zurich (Switzerland), and University of Washington, Seattle (USA) - whereby teachers and students, obviously on three different continents and in three different time zones, roughly eight hours apart, were working on a common design project using computer-aided design systems, video-conferencing and a web-based central database that managed and displayed all works throughout the process. The 24 hour design cycle is a metaphor for a more open and international approach to design, facilitated through computer networks. It implies a new form of collective authorship and distributed credits and thus deals with some of the essential challenges and opportunities the internet poses to creative disciplines.
series CAADRIA
email
last changed 2022/06/07 07:50

_id 70e3
authors Kim, Yong-Seong
year 1999
title Knowledge-Aided Design System for Intelligent Building Design
source CAADRIA '99 [Proceedings of The Fourth Conference on Computer Aided Architectural Design Research in Asia / ISBN 7-5439-1233-3] Shanghai (China) 5-7 May 1999, pp. 305-312
doi https://doi.org/10.52842/conf.caadria.1999.305
summary In the age of information technology, architectural design problems become increasingly complex, the finding of optimal solutions has become more difficult and obscure. Computer-aided design techniques have been applied to solve these ill-structured design problems; however, most of these applications have been used for graphical automation. Design improvement in quality has not been achieved using traditional computer programs. To handle the critical design decision problems, design systems need to be structured based on theoretical problem solving models. This would enable the design system to handle the problem solving design knowledge as well as the various technological aspects and geometrical representations. A theoretical model, knowledge-aided design, is proposed. Knowledge-aided design is a conceptual and theoretical model based on fundamental principles of design. It provides a problem-solving environment and a procedure for knowledge-based computer-aided architectural design based on cognitive science and artificial intelligence techniques. As a partial implementation of the theoretical model, the development of knowledge-aided design system for intelligent building design is described.
series CAADRIA
last changed 2022/06/07 07:52

_id dfa1
authors Laiserin, Jerry
year 1999
title CAD in Practice Profile: R.M. Kliment and Frances Halsband Architects
source ACADIA Quarterly, vol. 18, no. 1, pp. 15-18
doi https://doi.org/10.52842/conf.acadia.1999.015.2
summary R.M. Kliment and Frances Halsband Architects (KHA) is a firm recognized—among many outstanding achievements— for designing award-winning computer science centers at major universities (e.g., Columbia, Dartmouth, Princeton). With that design experience, it is no surprise that the firm has adopted an aggressive stance towards its own use of information technology (IT). One indication of this proactive approach to technology is that KHA, with a total staff of 33, carries a full-time CAD/systems manager position, as contrasted with the A/E-firm industry-wide average of one such full-time equivalent staff position for every 40 total employees. In effect, the firm has set its investment in and commitment to the role of IT at a rate twenty percent higher than the industry average. Such above-average investment in IT is consistent with other high-profile design firms that have won the prestigious Firm Award of the American Institute of Architects. (1)
series ACADIA
email
last changed 2022/06/07 07:52

_id ga0010
id ga0010
authors Moroni, A., Zuben, F. Von and Manzolli, J.
year 2000
title ArTbitrariness in Music
source International Conference on Generative Art
summary Evolution is now considered not only powerful enough to bring about the biological entities as complex as humans and conciousness, but also useful in simulation to create algorithms and structures of higher levels of complexity than could easily be built by design. In the context of artistic domains, the process of human-machine interaction is analyzed as a good framework to explore creativity and to produce results that could not be obtained without this interaction. When evolutionary computation and other computational intelligence methodologies are involved, every attempt to improve aesthetic judgement we denote as ArTbitrariness, and is interpreted as an interactive iterative optimization process. ArTbitrariness is also suggested as an effective way to produce art through an efficient manipulation of information and a proper use of computational creativity to increase the complexity of the results without neglecting the aesthetic aspects [Moroni et al., 2000]. Our emphasis will be in an approach to interactive music composition. The problem of computer generation of musical material has received extensive attention and a subclass of the field of algorithmic composition includes those applications which use the computer as something in between an instrument, in which a user "plays" through the application's interface, and a compositional aid, which a user experiments with in order to generate stimulating and varying musical material. This approach was adopted in Vox Populi, a hybrid made up of an instrument and a compositional environment. Differently from other systems found in genetic algorithms or evolutionary computation, in which people have to listen to and judge the musical items, Vox Populi uses the computer and the mouse as real-time music controllers, acting as a new interactive computer-based musical instrument. The interface is designed to be flexible for the user to modify the music being generated. It explores evolutionary computation in the context of algorithmic composition and provides a graphical interface that allows to modify the tonal center and the voice range, changing the evolution of the music by using the mouse[Moroni et al., 1999]. A piece of music consists of several sets of musical material manipulated and exposed to the listener, for example pitches, harmonies, rhythms, timbres, etc. They are composed of a finite number of elements and basically, the aim of a composer is to organize those elements in an esthetic way. Modeling a piece as a dynamic system implies a view in which the composer draws trajectories or orbits using the elements of each set [Manzolli, 1991]. Nonlinear iterative mappings are associated with interface controls. In the next page two examples of nonlinear iterative mappings with their resulting musical pieces are shown.The mappings may give rise to attractors, defined as geometric figures that represent the set of stationary states of a non-linear dynamic system, or simply trajectories to which the system is attracted. The relevance of this approach goes beyond music applications per se. Computer music systems that are built on the basis of a solid theory can be coherently embedded into multimedia environments. The richness and specialty of the music domain are likely to initiate new thinking and ideas, which will have an impact on areas such as knowledge representation and planning, and on the design of visual formalisms and human-computer interfaces in general. Above and bellow, Vox Populi interface is depicted, showing two nonlinear iterative mappings with their resulting musical pieces. References [Manzolli, 1991] J. Manzolli. Harmonic Strange Attractors, CEM BULLETIN, Vol. 2, No. 2, 4 -- 7, 1991. [Moroni et al., 1999] Moroni, J. Manzolli, F. Von Zuben, R. Gudwin. Evolutionary Computation applied to Algorithmic Composition, Proceedings of CEC99 - IEEE International Conference on Evolutionary Computation, Washington D. C., p. 807 -- 811,1999. [Moroni et al., 2000] Moroni, A., Von Zuben, F. and Manzolli, J. ArTbitration, Las Vegas, USA: Proceedings of the 2000 Genetic and Evolutionary Computation Conference Workshop Program – GECCO, 143 -- 145, 2000.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id d8df
authors Naticchia, Berardo
year 1999
title Physical Knowledge in Patterns: Bayesian Network Models for Preliminary Design
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 611-619
doi https://doi.org/10.52842/conf.ecaade.1999.611
summary Computer applications in design have pursued two main development directions: analytical modelling and information technology. The former line has produced a large number of tools for reality simulation (i.e. finite element models), the latter is producing an equally large amount of advances in conceptual design support (i.e. artificial intelligence tools). Nevertheless we can trace rare interactions between computation models related to those different approaches. This lack of integration is the main reason of the difficulty of CAAD application to the preliminary stage of design, where logical and quantitative reasoning are closely related in a process that we often call 'qualitative evaluation'. This paper briefly surveys the current development of qualitative physical models applied in design and propose a general approach for modelling physical behaviour by means of Bayesian network we are employing to develop a tutoring and coaching system for natural ventilation preliminary design of halls, called VENTPad. This tool explores the possibility of modelling the causal mechanism that operate in real systems in order to allow a number of integrated logical and quantitative inference about the fluid-dynamic behaviour of an hall. This application could be an interesting connection tool between logical and analytical procedures in preliminary design aiding, able to help students or unskilled architects, both to guide them through the analysis process of numerical data (i.e. obtained with sophisticate Computational Fluid Dynamics software) or experimental data (i.e. obtained with laboratory test models) and to suggest improvements to the design.
keywords Qualitative Physical Modelling, Preliminary Design, Bayesian Networks
series eCAADe
email
last changed 2022/06/07 07:59

_id a4e9
authors Petrovic, Igor and Svetel, Igor
year 1999
title From Number Cruncher to Digital Being: The Changing Role of Computer in CAAD
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 33-39
doi https://doi.org/10.52842/conf.ecaade.1999.033
summary The paper reflects on a thirteen-year period of CAAD research and development by a small group of researchers and practitioners. Starting with simple algorithmic drafting programmes, the work transcended to expert systems and distributed artificial intelligence, using computers as tools. The research cycle is about to begin afresh; computers in the next century shall not be detached entities but the extensions of man. The computer shall be the medium that will enable a designer to be what he/she really is. This future has already begun.
keywords History of CAAD, CAAD Design Paradigms, CAADfuture
series eCAADe
email
last changed 2022/06/07 08:00

_id ga9908
id ga9908
authors Senagala, Mahesh
year 1999
title Artistic Process, Cybernetics of Self and the Epistemology of Digital Technology
source International Conference on Generative Art
summary From the viewpoint of Batesonian cybernetics, ‘conscious purpose’ and artistic process are distinct ends of a spectrum of the functioning of self. Artistic activities— by which I mean art, poetry, play, design, etc.— involve processes that are beneath the stratum of consciousness. By definition, consciousness is selective awareness and is linear in execution and limited in its capability to synthesize complex parameters. As Heidegger pointed out, technology is a special form of knowledge (episteme). A machine is a manifestation of such a knowledge. A machine is a result of conscious purpose and is normally task-driven to accomplish a specific purpose(s). The questions this paper raises are to do with the connections between conscious purpose, artistic process and digital technology. One of the central questions of the paper is "if artistic process requires an abandonment or relinquishment of conscious purpose at the time of the generation of the work of art, and if the artistic process is a result of vast number of ‘unconscious’ forces and impulses, then could we say that the computer would ever be able to ‘generate’ or ‘create’ a work of art?" In what capacity and what role would the computer be a part of the generative process of art? Would a computer be able to ‘generate’ and ‘know’ a work of art, which, according to Bateson, requires the abandonment of conscious purpose? The ultimate goal of the paper is to unearth and examine the potential of the computers to be a part of the generative process of what Bateson has called "total self as a cybernetic model". On another plane of discourse, Deleuze and Guattari have added a critical dimension to the discourse of cybernetics and models of human mind and the global computer networks. Their notion of ‘rhizome’ has its roots in Batesonian cybernetics and the cybernetic couplings between the ‘complex systems’ such as human mind, biological and computational systems. Deleuze and Guattari call such systems as human brain and the neural networks as rhizomatic. Given the fact that the computer is the first known cybernetic machine to lay claims to artificial intelligence, the aforementioned questions become even more significant. The paper will explore how, cybernetically, the computer could be ‘coupled’ with ‘self’ and the artistic process — the ultimate expression of human condition. These philosophical and artistic explorations will take place through a series of generative artistic projects (See the figure below for an example) that aim at understanding the couplings and ‘ecology’ of digital technology and the cybernetics of self.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id c041
authors Vakalo, E., Malkawi, A.M. and Emdanat, S.S.
year 1999
title An AI-based shell for linking thermal and form-making considerations
source Automation in Construction 8 (4) (1999) pp. 455-462
summary Over the past few years, our team has developed several computer-based models in the areas of architectural form-making and thermal analysis. These programs were designed to deal with specific problems and use a range of techniques including machine vision, knowledge-based systems, and artificial intelligence techniques. Recently, a project that integrates these systems was initiated. Its objective is to design an intelligent computer shell that forms the basis for this integration in the domain of architecture. The paper discusses the development of the shell and its use to analyze and study architectural form and its determinants. The shell accommodates modules that link the morphological structure of architectural design with more of its determinants (e.g., structural, acoustical, and lighting considerations, as well as code requirements). The paper presents and discusses the background of the shell, its structure, its methods of knowledge representation, and an example of its use.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:23

_id avocaad_2001_22
id avocaad_2001_22
authors Jos van Leeuwen, Joran Jessurun
year 2001
title XML for Flexibility an Extensibility of Design Information Models
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary The VR-DIS research programme aims at the development of a Virtual Reality – Design Information System. This is a design and decision support system for collaborative design that provides a VR interface for the interaction with both the geometric representation of a design and the non-geometric information concerning the design throughout the design process. The major part of the research programme focuses on early stages of design. The programme is carried out by a large number of researchers from a variety of disciplines in the domain of construction and architecture, including architectural design, building physics, structural design, construction management, etc.Management of design information is at the core of this design and decision support system. Much effort in the development of the system has been and still is dedicated to the underlying theory for information management and its implementation in an Application Programming Interface (API) that the various modules of the system use. The theory is based on a so-called Feature-based modelling approach and is described in the PhD thesis by [first author, 1999] and in [first author et al., 2000a]. This information modelling approach provides three major capabilities: (1) it allows for extensibility of conceptual schemas, which is used to enable a designer to define new typologies to model with; (2) it supports sharing of conceptual schemas, called type-libraries; and (3) it provides a high level of flexibility that offers the designer the opportunity to easily reuse design information and to model information constructs that are not foreseen in any existing typologies. The latter aspect involves the capability to expand information entities in a model with relationships and properties that are not typologically defined but applicable to a particular design situation only; this helps the designer to represent the actual design concepts more accurately.The functional design of the information modelling system is based on a three-layered framework. In the bottom layer, the actual design data is stored in so-called Feature Instances. The middle layer defines the typologies of these instances in so-called Feature Types. The top layer is called the meta-layer because it provides the class definitions for both the Types layer and the Instances layer; both Feature Types and Feature Instances are objects of the classes defined in the top layer. This top layer ensures that types can be defined on the fly and that instances can be created from these types, as well as expanded with non-typological properties and relationships while still conforming to the information structures laid out in the meta-layer.The VR-DIS system consists of a growing number of modules for different kinds of functionality in relation with the design task. These modules access the design information through the API that implements the meta-layer of the framework. This API has previously been implemented using an Object-Oriented Database (OODB), but this implementation had a number of disadvantages. The dependency of the OODB, a commercial software library, was considered the most problematic. Not only are licenses of the OODB library rather expensive, also the fact that this library is not common technology that can easily be shared among a wide range of applications, including existing applications, reduces its suitability for a system with the aforementioned specifications. In addition, the OODB approach required a relatively large effort to implement the desired functionality. It lacked adequate support to generate unique identifications for worldwide information sources that were understandable for human interpretation. This strongly limited the capabilities of the system to share conceptual schemas.The approach that is currently being implemented for the core of the VR-DIS system is based on eXtensible Markup Language (XML). Rather than implementing the meta-layer of the framework into classes of Feature Types and Feature Instances, this level of meta-definitions is provided in a document type definition (DTD). The DTD is complemented with a set of rules that are implemented into a parser API, based on the Document Object Model (DOM). The advantages of the XML approach for the modelling framework are immediate. Type-libraries distributed through Internet are now supported through the mechanisms of namespaces and XLink. The implementation of the API is no longer dependent of a particular database system. This provides much more flexibility in the implementation of the various modules of the VR-DIS system. Being based on the (supposed to become) standard of XML the implementation is much more versatile in its future usage, specifically in a distributed, Internet-based environment.These immediate advantages of the XML approach opened the door to a wide range of applications that are and will be developed on top of the VR-DIS core. Examples of these are the VR-based 3D sketching module [VR-DIS ref., 2000]; the VR-based information-modelling tool that allows the management and manipulation of information models for design in a VR environment [VR-DIS ref., 2000]; and a design-knowledge capturing module that is now under development [first author et al., 2000a and 2000b]. The latter module aims to assist the designer in the recognition and utilisation of existing and new typologies in a design situation. The replacement of the OODB implementation of the API by the XML implementation enables these modules to use distributed Feature databases through Internet, without many changes to their own code, and without the loss of the flexibility and extensibility of conceptual schemas that are implemented as part of the API. Research in the near future will result in Internet-based applications that support designers in the utilisation of distributed libraries of product-information, design-knowledge, case-bases, etc.The paper roughly follows the outline of the abstract, starting with an introduction to the VR-DIS project, its objectives, and the developed theory of the Feature-modelling framework that forms the core of it. It briefly discusses the necessity of schema evolution, flexibility and extensibility of conceptual schemas, and how these capabilities have been addressed in the framework. The major part of the paper describes how the previously mentioned aspects of the framework are implemented in the XML-based approach, providing details on the so-called meta-layer, its definition in the DTD, and the parser rules that complement it. The impact of the XML approach on the functionality of the VR-DIS modules and the system as a whole is demonstrated by a discussion of these modules and scenarios of their usage for design tasks. The paper is concluded with an overview of future work on the sharing of Internet-based design information and design knowledge.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id 4a1a
authors Laird, J.E.
year 2001
title Using Computer Game to Develop Advanced AI
source Computer, 34 (7), July pp. 70-75
summary Although computer and video games have existed for fewer than 40 years, they are already serious business. Entertainment software, the entertainment industry's fastest growing segment, currently generates sales surpassing the film industry's gross revenues. Computer games have significantly affected personal computer sales, providing the initial application for CD-ROMs, driving advancements in graphics technology, and motivating the purchase of ever faster machines. Next-generation computer game consoles are extending this trend, with Sony and Toshiba spending $2 billion to develop the Playstation 2 and Microsoft planning to spend more than $500 million just to market its Xbox console [1]. These investments have paid off. In the past five years, the quality and complexity of computer games have advanced significantly. Computer graphics have shown the most noticeable improvement, with the number of polygons rendered in a scene increasing almost exponentially each year, significantly enhancing the games' realism. For example, the original Playstation, released in 1995, renders 300,000 polygons per second, while Sega's Dreamcast, released in 1999, renders 3 million polygons per second. The Playstation 2 sets the current standard, rendering 66 million polygons per second, while projections indicate the Xbox will render more than lOO million polygons per second. Thus, the images on today's $300 game consoles rival or surpass those available on the previous decade's $50,000 computers. The impact of these improvements is evident in the complexity and realism of the environments underlying today's games, from detailed indoor rooms and corridors to vast outdoor landscapes. These games populate the environments with both human and computer controlled characters, making them a rich laboratory for artificial intelligence research into developing intelligent and social autonomous agents. Indeed, computer games offer a fitting subject for serious academic study, undergraduate education, and graduate student and faculty research. Creating and efficiently rendering these environments touches on every topic in a computer science curriculum. The "Teaching Game Design " sidebar describes the benefits and challenges of developing computer game design courses, an increasingly popular field of study
series journal paper
last changed 2003/04/23 15:50

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 30HOMELOGIN (you are user _anon_271527 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002