CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 626

_id ga9921
id ga9921
authors Coates, P.S. and Hazarika, L.
year 1999
title The use of genetic programming for applications in the field of spatial composition
source International Conference on Generative Art
summary Architectural design teaching using computers has been a preoccupation of CECA since 1991. All design tutors provide their students with a set of models and ways to form, and we have explored a set of approaches including cellular automata, genetic programming ,agent based modelling and shape grammars as additional tools with which to explore architectural ( and architectonic) ideas.This paper discusses the use of genetic programming (G.P.) for applications in the field of spatial composition. CECA has been developing the use of Genetic Programming for some time ( see references ) and has covered the evolution of L-Systems production rules( coates 1997, 1999b), and the evolution of generative grammars of form (Coates 1998 1999a). The G.P. was used to generate three-dimensional spatial forms from a set of geometrical structures .The approach uses genetic programming with a Genetic Library (G.Lib) .G.P. provides a way to genetically breed a computer program to solve a problem.G. Lib. enables genetic programming to define potentially useful subroutines dynamically during a run .* Exploring a shape grammar consisting of simple solid primitives and transformations. * Applying a simple fitness function to the solid breeding G.P.* Exploring a shape grammar of composite surface objects. * Developing grammarsfor existing buildings, and creating hybrids. * Exploring the shape grammar of abuilding within a G.P.We will report on new work using a range of different morphologies ( boolean operations, surface operations and grammars of style ) and describe the use of objective functions ( natural selection) and the "eyeball test" ( artificial selection) as ways of controlling and exploring the design spaces thus defined.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id ae38
authors Jabi, Wassim
year 1999
title Integrating Databases, Objects and the World-Wide Web for Collaboration in Architectural Design
source Proceedings of the focus symposium: World Wide Web as Framework for Collaboration in conjunction with the 11th International Conference on Systems Research, Informatics and Cybernetics, The International Institute for Advanced Studies in Systems Research
summary Architectural design requires specialized vertical knowledge that goes beyond the sharing of marks on paper or the multi-casting of video images. This paper briefly surveys the state-ofthe- art in groupware and outlines the need for vertical and integrated support of synchronous and asynchronous design collaboration. The paper also describes a software prototype (WebOutliner) under development that uses a three-tier persistent object-oriented, web-based technology for a richer representation of hierarchical architectural artifacts using Apple’s WebObjects technology. The prototype contributes to earlier work that defined a framework for a shared workspace consisting of Participants, Tasks, Proposals, and Artifacts. These four elements have been found through observation and analysis to be adequate representations of the essential components of collaborative architectural design. These components are also hierarchical which allows users to filter information, copy completed solutions to other parts of the program, analyze and compare design parameters and aggregate hierarchical amounts. Given its object orientation, the represented artifacts have built-in data and methods that allow them to respond to user actions and manage their own sub-artifacts. In addition, the prototype integrates this technology with Java tools for ubiquitous synchronous web-based access. The prototype uses architectural programming (defining the spatial program of a building) and early conceptual design as examples of seamlessly integrated groupware applications.
keywords Computer Supported Collaborative Design, WebObjects, Synchronous and Asynchronous Collaboration, Java Applets, Application Server, Web-based Interface
series other
email
last changed 2002/03/05 19:55

_id avocaad_2001_22
id avocaad_2001_22
authors Jos van Leeuwen, Joran Jessurun
year 2001
title XML for Flexibility an Extensibility of Design Information Models
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary The VR-DIS research programme aims at the development of a Virtual Reality – Design Information System. This is a design and decision support system for collaborative design that provides a VR interface for the interaction with both the geometric representation of a design and the non-geometric information concerning the design throughout the design process. The major part of the research programme focuses on early stages of design. The programme is carried out by a large number of researchers from a variety of disciplines in the domain of construction and architecture, including architectural design, building physics, structural design, construction management, etc.Management of design information is at the core of this design and decision support system. Much effort in the development of the system has been and still is dedicated to the underlying theory for information management and its implementation in an Application Programming Interface (API) that the various modules of the system use. The theory is based on a so-called Feature-based modelling approach and is described in the PhD thesis by [first author, 1999] and in [first author et al., 2000a]. This information modelling approach provides three major capabilities: (1) it allows for extensibility of conceptual schemas, which is used to enable a designer to define new typologies to model with; (2) it supports sharing of conceptual schemas, called type-libraries; and (3) it provides a high level of flexibility that offers the designer the opportunity to easily reuse design information and to model information constructs that are not foreseen in any existing typologies. The latter aspect involves the capability to expand information entities in a model with relationships and properties that are not typologically defined but applicable to a particular design situation only; this helps the designer to represent the actual design concepts more accurately.The functional design of the information modelling system is based on a three-layered framework. In the bottom layer, the actual design data is stored in so-called Feature Instances. The middle layer defines the typologies of these instances in so-called Feature Types. The top layer is called the meta-layer because it provides the class definitions for both the Types layer and the Instances layer; both Feature Types and Feature Instances are objects of the classes defined in the top layer. This top layer ensures that types can be defined on the fly and that instances can be created from these types, as well as expanded with non-typological properties and relationships while still conforming to the information structures laid out in the meta-layer.The VR-DIS system consists of a growing number of modules for different kinds of functionality in relation with the design task. These modules access the design information through the API that implements the meta-layer of the framework. This API has previously been implemented using an Object-Oriented Database (OODB), but this implementation had a number of disadvantages. The dependency of the OODB, a commercial software library, was considered the most problematic. Not only are licenses of the OODB library rather expensive, also the fact that this library is not common technology that can easily be shared among a wide range of applications, including existing applications, reduces its suitability for a system with the aforementioned specifications. In addition, the OODB approach required a relatively large effort to implement the desired functionality. It lacked adequate support to generate unique identifications for worldwide information sources that were understandable for human interpretation. This strongly limited the capabilities of the system to share conceptual schemas.The approach that is currently being implemented for the core of the VR-DIS system is based on eXtensible Markup Language (XML). Rather than implementing the meta-layer of the framework into classes of Feature Types and Feature Instances, this level of meta-definitions is provided in a document type definition (DTD). The DTD is complemented with a set of rules that are implemented into a parser API, based on the Document Object Model (DOM). The advantages of the XML approach for the modelling framework are immediate. Type-libraries distributed through Internet are now supported through the mechanisms of namespaces and XLink. The implementation of the API is no longer dependent of a particular database system. This provides much more flexibility in the implementation of the various modules of the VR-DIS system. Being based on the (supposed to become) standard of XML the implementation is much more versatile in its future usage, specifically in a distributed, Internet-based environment.These immediate advantages of the XML approach opened the door to a wide range of applications that are and will be developed on top of the VR-DIS core. Examples of these are the VR-based 3D sketching module [VR-DIS ref., 2000]; the VR-based information-modelling tool that allows the management and manipulation of information models for design in a VR environment [VR-DIS ref., 2000]; and a design-knowledge capturing module that is now under development [first author et al., 2000a and 2000b]. The latter module aims to assist the designer in the recognition and utilisation of existing and new typologies in a design situation. The replacement of the OODB implementation of the API by the XML implementation enables these modules to use distributed Feature databases through Internet, without many changes to their own code, and without the loss of the flexibility and extensibility of conceptual schemas that are implemented as part of the API. Research in the near future will result in Internet-based applications that support designers in the utilisation of distributed libraries of product-information, design-knowledge, case-bases, etc.The paper roughly follows the outline of the abstract, starting with an introduction to the VR-DIS project, its objectives, and the developed theory of the Feature-modelling framework that forms the core of it. It briefly discusses the necessity of schema evolution, flexibility and extensibility of conceptual schemas, and how these capabilities have been addressed in the framework. The major part of the paper describes how the previously mentioned aspects of the framework are implemented in the XML-based approach, providing details on the so-called meta-layer, its definition in the DTD, and the parser rules that complement it. The impact of the XML approach on the functionality of the VR-DIS modules and the system as a whole is demonstrated by a discussion of these modules and scenarios of their usage for design tasks. The paper is concluded with an overview of future work on the sharing of Internet-based design information and design knowledge.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id 04fd
authors Karhu, V. and Lahdenpera, P.
year 1999
title A formalised process model of current Finnish design and construction practice
source The Int. Journal of Construction IT 7(1), pp. 51-71
summary There is a need for improved co-ordination to enhance the performance of the building process. The process involves many parties and the communication and interfaces need special attention. Conventionally, the processes of parties are carried out independently, each discipline having its own activities and limits. As a precursor to improving the overall process, formal process modelling may be used to clarify the activities, information flows and the responsibilities of the different parties. The model presented in this paper divides the Finnish construction process into six main stages: briefing, programming, global design, detailed design, construction and hand-over. In developing the model, all these stages were covered - the main focus being on the functions and flows of the process since these were found to be the most critical in the development of the building procedures. The IDEF0 method was used as the modelling technique. It is shown how the developed reference model can be subjected to various view-dependent examinations and that the modelling approach supports process re-engineering and improvement efforts as well as a new means of building process management, especially when combined with modern computer-aided applications.
series journal paper
last changed 2003/05/15 21:45

_id 39cb
authors Kelleners, Richard H.M.C.
year 1999
title Constraints in object-oriented graphics
source Eindhoven University of Technology
summary In the area of interactive computer graphics, two important approaches to deal with the complexity of designing and implementing graphics systems are object-oriented programming and constraint-based programming. From literature, it appears that combination of these two has clear advantages but has also proven to be difficult. One of the main problems is that constraint programming infringes the information hiding principle of object-oriented programming. The goal of the research project is to combine these two approaches to benefit from the strengths of both. Two research groups at the Eindhoven University of Technology investigate the use of constraints on graphics objects. At the Architecture department, constraints are applied in a virtual reality design environment. At the Computer Science department, constraints aid in modeling 3D animations. For these two groups, a constraint system for 3D graphical objects was developed. A conceptual model, called CODE (Constraints on Objects via Data flows and Events), is presented that enables integration of constraints and objects by separating the object world from the constraint world. In the design of this model, the main aspect being considered is that the information hiding principle among objects may not be violated. Constraint solvers, however, should have direct access to an object’s internal data structure. Communication between the two worlds is done via a protocol orthogonal to the message passing mechanism of objects, namely, via events and data flows. This protocol ensures that the information hiding principle at the object-oriented programming level is not violated while constraints can directly access “hidden” data. Furthermore, CODE is built up of distinct elements, or entity types, like constraint, solver, event, data flow. This structure enables that several special purpose constraint solvers can be defined and made to cooperate to solve complex constraint problems. A prototype implementation was built to study the feasibility of CODE. Therefore, the implementation should correspond directly to the conceptual model. To this end, every entity (object, constraint, solver) of the conceptual model is represented by a separate process in the language MANIFOLD. The (concurrent) processes communicate by events and data flows. The implementation serves to validate the conceptual model and to demonstrate that it is a viable way of combining constraints and objects. After the feasibility study, the prototype was discarded. The gained experience was used to build an implementation of the conceptual model for the two research groups. This implementation encompassed a constraint system with multiple solvers and constraint types. The constraint system was built as an object-oriented library that can be linked to the applications in the respective research groups. Special constructs were designed to ensure information hiding among application objects while constraints and solvers have direct access to the object data. CODE manages the complexity of object-oriented constraint solving by defining a communication protocol to allow the two paradigms to cooperate. The prototype implementation demonstrates that CODE can be implemented into a working system. Finally, the implementation of an actual application shows that the model is suitable for the development of object-oriented software.
keywords Computer Graphics; Object Oriented Programming; Constraint Programming
series thesis:PhD
last changed 2003/02/12 22:37

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id d81d
authors Pierce, Jeffrey S., Conway, M., van Dantzich, M. and Robertson, G.
year 1999
title Toolspaces and Glances: Storing, Accessing, and Retrieving Objects in 3D Desktop Applications
source Proceedings of the Symposium on Interactive 3D Graphics 1999. pp. 163-168
summary Users of 3D desktop applications perform tasks that require accessing data storage, moving objects, and navigation. These operations are typically performed using 2D GUI elements or 3D widgets. We wish to focus on interaction with 3D widgets directly in the 3D world, rather than forcing our users to repeatedly switch contexts between 2D and 3D. However, the use of 3D widgets requires a mechanism for storing, accessing, and retrieving these widgets. In this paper we present toolspaces and glances to provide this capability for 3D widgets and other objects in interactive 3D worlds. Toolspaces are storage spaces attached to the user's virtual body; objects placed in these spaces are always accessible yet out of the user's view until needed. Users access these toolspaces to store and retrieve objects through a type of lightweight and ephemeral navigation we call glances.
series other
last changed 2003/04/23 15:50

_id 1570
authors Sowizral, H.A. and Deering, M.F.
year 1999
title The Java 3D API and Virtual Reality
source IEEE Computer Graphics and Applications, May/June
summary Java 3D proves a natural choice for any Java programmer wanting to write an interactive 3D graphics program. A programmer constructs a scene graph containing graphic objects, lights, sounds, environmental effects objects, and behavior objects that handle interactions or modify other objects in the scene graph. The programmer then hands that scene graph to Java 3D for execution. Java 3D starts rendering objects and executing behaviors in the scene graph. Virtual reality applications go through an identical writing process. However, before a user can use such an application, Java 3D must additionally know about the user's physical characteristics (height, eye separation, and so forth) and physical environment (number of displays, their location, trackers, and so on). Not surprisingly, such information varies from installation to installation and from user to user. So Java 3D lets application developers separate their application's operation from the vagaries of the user's final display environment. The Java 3D application programmer's interface (API) provides a very flexible platform for building a broad range of graphics applications. Developers have already used Java 3D to build applications in a variety of domains including mechanical CAD, molecular visualization, scientific visualization, animation previews, geographic information systems, business graphics, 3D logos, and educational offerings. Virtual reality applications have included immersive workbench applications, headtracked shutter-glass-based desktop applications, and portals (a cave-like room with multiple back-projected walls).
series journal paper
last changed 2003/04/23 15:50

_id 21c8
authors Bargiela, Beatriz and Bausset, Raúl Abad
year 1999
title Sistemas multimediales aplicados a la arquitectura y su conocimiento (Multimedia Systems Applied to Architecture and its Knowledge Base)
source III Congreso Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings] Montevideo (Uruguay) September 29th - October 1st 1999, pp. 186-191
summary This text makes a careful description of a great part of the components of the vocabulary of architecture and their syntax. It's structured in two parts, one textual and the other graphic. Due to the acknowledged contribution of the author on the subject of description of vocabulary of the elements of architecture and the research and development works that are entered upon in the master in computer graphics in the informatic field, it was proposed the development of an interactive system that allowed by means of links to connect the different parts of text and their respective graphics. Having as a basis the idiomatic equivalent term in Spanish and English, carried out in another research work, it was considered the possibility to link the text in these two languages besides the original one, french. That's why we've decided to propose the navigation through numeric text and images. This navigation has been already inferred from the reading of the text as intended from the author's that was why our task has been the interpretation of this intention and its translation in an informatic system. The product presented in this work is limited to having designed a methodology and showing its performance with some terms and images, able to make evident the sketch of our idea. The work is made out of the description of the following phases: 1.) Design of its interface; 2.) Compilation and classification of images compilation and classification of text; 3.) Coordination of textual and graphic elements; 4.) Programming of events in the interactive system.
series SIGRADI
email
last changed 2016/03/10 09:47

_id 9a1e
authors Clayton, Mark J. and Vasquez de Velasco, Guillermo
year 1999
title Stumbling, Backtracking, and Leapfrogging: Two Decades of Introductory Architectural Computing
doi https://doi.org/10.52842/conf.ecaade.1999.151
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 151-158
summary Our collective concept of computing and its relevance to architecture has undergone dramatic shifts in emphasis. A review of popular texts from the past reveals the biases and emphases that were current. In the seventies, architectural computing was generally seen as an elective for data processing specialists. In the early eighties, personal computers and commercial CAD systems were widely adopted. Architectural computing diverged from the "batch" world into the "interactive" world. As personal computing matured, introductory architectural computing courses turned away from a foundation in programming toward instruction in CAD software. By the late eighties, Graphic User Interfaces and windowing operating systems had appeared, leading to a profusion of architecturally relevant applications that needed to be addressed in introductory computing. The introduction of desktop 3D modeling in the early nineties led to increased emphasis upon rendering and animation. The past few years have added new emphases, particularly in the area of network communications, the World Wide Web and Virtual Design Studios. On the horizon are topics of electronic commerce and knowledge markets. This paper reviews these past and current trends and presents an outline for an introductory computing course that is relevant to the year 2000.
keywords Computer-Aided Architectural Design, Computer-Aided Design, Computing Education, Introductory Courses
series eCAADe
email
last changed 2022/06/07 07:56

_id 99ce
authors Forowicz, T.
year 1999
title Modeling of energy demands for residential buildings with HTML interface
source Automation in Construction 8 (4) (1999) pp. 481-487
summary This paper presents the package for calculation of energy and cost demands for heating, cooling and hot water. The package represents a new kind of approach to developing software, employing user (client) and server (program provider) computers connected by Internet. It is mounted on the owner server and is available to the whole world through the Web browser. The package was developed as a simplified tool for estimating energy use in four types of new and old houses, located in 900 US cities. The computing engine utilizes the database that was compiled by LBL in support of the 'Affordable Housing through Energy Conservation' Project with over 10000 DOE-2.1 simulations. The package consists of 69 routines and scripts coded in four languages: HTML, Perl, C, and FORTRAN. The modeling, the programming, and the future perspectives of the new kind of computational tool are presented. The paper discusses further technical limitations, as well as suggestions for further improvements and development. Especially important is the problem of multi-user access; ways for its solution are proposed.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:22

_id 0a45
authors Lee, K.
year 1999
title Principles of CAD/CAM/CAE Systems
source New York: Addison-Wesley 1999
summary An Introduction to CAD/CAM/CAE is a timely text with coverage of many modern topics, including: rapid prototyping, virtual engineering, NT-based solid modeling systems, and Web-related issues. This book provides balanced coverage of CAD/CAM (with slightly more emphasis on CAD topics) and bonus coverage of computer-aided engineering (CAE). This book's emphasis on the integration of three related disciplines-CAD, CAM, and CAE-makes it a great theoretical introduction to all subjects from geometric representation to the most sophisticated CAE subjects. Many illustrations and references ground the theory in practical examples. The book also features a unique illustration of the whole product development process through a practical case study that gives readers a clear idea how CAD, CAM, and CAE systems are integrated to accelerate the product development process. This book is carefully targeted toward today's students; topics are introduced in a concise, efficient manner, with mathematical terminology that is kept to a minimum. A background in programming, calculus, and matrix and vector algebra is helpful when using this text.
series other
last changed 2003/04/23 15:14

_id ga0009
id ga0009
authors Lewis, Matthew
year 2000
title Aesthetic Evolutionary Design with Data Flow Networks
source International Conference on Generative Art
summary For a little over a decade, software has been created which allows for the design of visual content by aesthetic evolutionary design (AED) [3]. The great majority of these AED systems involve custom software intended for breeding entities within one fairly narrow problem domain, e.g., certain classes of buildings, cars, images, etc. [5]. Only a very few generic AED systems have been attempted, and extending them to a new design problem domain can require a significant amount of custom software development [6][8]. High end computer graphics software packages have in recent years become sufficiently robust to allow for flexible specification and construction of high level procedural models. These packages also provide extensibility, allowing for the creation of new software tools. One component of these systems which enables rapid development of new generative models and tools is the visual data flow network [1][2][7]. One of the first CG packages to employ this paradigm was Houdini. A system constructed within Houdini which allows for very fast generic specification of evolvable parametric prototypes is described [4]. The real-time nature of the software, when combined with the interlocking data networks, allows not only for vertical ancestor/child populations within the design space to be explored, but also allows for fast "horizontal" exploration of the potential population surface. Several example problem domains will be presented and discussed. References: [1] Alias | Wavefront. Maya. 2000, http://www.aliaswavefront.com [2] Avid. SOFTIMAGE. 2000, http://www.softimage.com [3] Bentley, Peter J. Evolutionary Design by Computers. Morgan Kaufmann, 1999. [4] Lewis, Matthew. "Metavolve Home Page". 2000, http://www.cgrg.ohio-state.edu/~mlewis/AED/Metavolve/ [5] Lewis, Matthew. "Visual Aesthetic Evolutionary Design Links". 2000, http://www.cgrg.ohio-state.edu/~mlewis/aed.html [6] Rowley, Timothy. "A Toolkit for Visual Genetic Programming". Technical Report GCG-74, The Geometry Center, University of Minnesota, 1994. [7] Side Effects Software. Houdini. 2000, http://www.sidefx.com [8] Todd, Stephen and William Latham. "The Mutation and Growth of Art by Computers" in Evolutionary Design by Computers, Peter Bentley ed., pp. 221-250, Chapter 9, Morgan Kaufmann, 1999.    
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 6810
authors Makkonen, Petri
year 1999
title On multi body systems simulation in product design
source KTH Stockholm
summary The aim of this thesis is to provide a basis for efficient modelling and software use in simulation driven product development. The capabilities of modern commercial computer software for design are analysed experimentally and qualitatively. An integrated simulation model for design of mechanical systems, based on four different "simulation views" is proposed: An integrated CAE (Computer Aided Engineering) model using Solid Geometry (CAD), Finite Element Modelling (FEM), Multi Body Systems Modelling (MBS) and Dynamic System Simulation utilising Block System Modelling tools is presented. A theoretical design process model for simulation driven design based on the theory of product chromosome is introduced. This thesis comprises a summary and six papers. Paper A presents the general framework and a distributed model for simulation based on CAD, FEM, MBS and Block Systems modelling. Paper B outlines a framework to integrate all these models into MBS simulation for performance prediction and optimisation of mechanical systems, using a modular approach. This methodology has been applied to design of industrial robots of parallel robot type. During the development process, from concept design to detail design, models have been refined from kinematic to dynamic and to elastodynamic models, finally including joint backlash. A method for analysing the kinematic Jacobian by using MBS simulation is presented. Motor torque requirements are studied by varying major robot geometry parameters, in dimensionless form for generality. The robot TCP (Tool Center Point) path in time space, predicted from elastodynamic model simulations, has been transformed to the frequency space by Fourier analysis. By comparison of this result with linear (modal) eigen frequency analysis from the elastodynamic MBS model, internal model validation is obtained. Paper C presents a study of joint backlash. An impact model for joint clearance, utilised in paper B, has been developed and compared to a simplified spring-damper model. The impact model was found to predict contact loss over a wider range of rotational speed than the spring-damper model. Increased joint bearing stiffness was found to widen the speed region of chaotic behaviour, due to loss of contact, while increased damping will reduce the chaotic range. The impact model was found to have stable under- and overcritical speed ranges, around the loss of contact region. The undercritical limit depends on the gravitational load on the clearance joint. Papers D and E give examples of the distributed simulation model approach proposed in paper A. Paper D presents simulation and optimisation of linear servo drives for a 3-axis gantry robot, using block systems modelling. The specified kinematic behaviour is simulated with multi body modelling, while drive systems and control system are modelled using a block system model for each drive. The block system model has been used for optimisation of the transmission and motor selection. Paper E presents an approach for re-using CAD geometry for multi body modelling of a rock drilling rig boom. Paper F presents synthesis methods for mechanical systems. Joint and part number synthesis is performed using the Grübler and Euler equations. The synthesis is continued by applying the theory of generative grammar, from which the grammatical rules of planar mechanisms have been formulated. An example of topological synthesis of mechanisms utilising this grammar is presented. Finally, dimensional synthesis of the mechanism is carried out by utilising non-linear programming with addition of a penalty function to avoid singularities.
keywords Simulation; Optimisation; Control Systems; Computer Aided Engineering; Multi Body Systems; Finite Element Method; Backslash; Clearance; Industrial Robots; Parallel Robots
series thesis:PhD
last changed 2003/02/12 22:37

_id 0c9c
authors Tweed, Christopher
year 1999
title Prescribing Designs
doi https://doi.org/10.52842/conf.ecaade.1999.051
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 51-57
summary Much of the debate and argument among CAAD researchers has turned on the degree to which CAAD systems limit the ways in which designers can express themselves. By defining representations for design objects and design functions, systems determine what it is possible to describe. Aart Bijl used the term 'prescriptiveness' to refer to this property of systems, and the need to overcome it was a major preoccupation of research at EdCAAD during the 1980s, including the development of the MOLE (Modelling Objects with Logic Expressions) system. But in trying to offer designers the freedom that was judged to be essential to evolving design practices, MOLE transferred much of the burden of programming from system developers to end-users - you can have any design objects you want, as long as you write the code. Close examination of MOLE's logic reveals that it too had to rely on fundamental definitions that, even if not domain-specific, are certainly historically contingent. This paper will return to the issue of prescriptiveness, summarising the lessons learned from the MOLE 'experiment,' and identifying new prescriptions that are deciding what designs can be. Looking beyond computer representations, we find that designs are shaped by much larger, and arguably more powerful, historical, social and cultural forces surrounding design practice. These forces are shaping the way CAAD is used and how new systems are conceived and developed.
keywords Bijl, Prescriptiveness
series eCAADe
email
last changed 2022/06/07 07:58

_id ga9903
id ga9903
authors Ward, Adrian and Cox, Geoff
year 1999
title How I Drew One of My Pictures: * or, The Authorship of Generative Art
source International Conference on Generative Art
summary The concept of value is traditionally bestowed on a work of art when it is seen to be unique and irreproducible, thereby granting it authenticity. Think of a famous painting: only the original canvas commands genuinely high prices. Digital artwork is not valued in the same way. It can be copied infinitely and there is therefore a corresponding crisis of value. It has been argued that under these conditions of the dematerialised artwork, it is process that becomes valued. In this way, the process of creation and creativity is valued in place of authenticity, undermining conventional notions of authorship. It is possible to correlate many of these creative processes into instructions. However, to give precise instructions on the construction of a creative work is a complex, authentic and intricate process equivalent to conventional creative work (and is therefore not simply a question of 'the death of the author'). This paper argues that to create ‘generative’ systems is a rigorous and intricate procedure. Moreover, the output from generative systems should not be valued simply as an endless, infinite series of resources but as a system. To have a machine write poetry for ten years would not generate creative music, but the process of getting the machine to do so would certainly register an advanced form of creativity. When a programmer develops a generative system, they are engaged in a creative act. Programming is no less an artform than painting is a technical process. By analogy, the mathematical value pi can be approximated as 3.14159265, but a more thorough and accurate version can be stored as the formula used to calculate it. In the same way, it is more complete to express creativity formulated as code, which can then be executed to produce the results we desire. Rather like using Leibnitz's set of symbols to represent a mathematical formula, artists can now choose to represent creativity as computer programs (Harold Cohen’s Aaron, a computer program that creates drawings is a case in point). By programming computers to undertake creative instructions, this paper will argue that more accurate and expansive traces of creativity are being developed that suitably merge artistic subjectivity with technical form. It is no longer necessary or even desirable to be able to render art as a final tangible medium, but instead it is more important to program computers to be creative by proxy. [The paper refers to Autoshop software, available from http://autoshop.signwave.co.uk]
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id cf2011_p109
id cf2011_p109
authors Abdelmohsen, Sherif; Lee Jinkook, Eastman Chuck
year 2011
title Automated Cost Analysis of Concept Design BIM Models
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 403-418.
summary AUTOMATED COST ANALYSIS OF CONCEPT DESIGN BIM MODELS Interoperability: BIM models and cost models This paper introduces the automated cost analysis developed for the General Services Administration (GSA) and the analysis results of a case study involving a concept design courthouse BIM model. The purpose of this study is to investigate interoperability issues related to integrating design and analysis tools; specifically BIM models and cost models. Previous efforts to generate cost estimates from BIM models have focused on developing two necessary but disjoint processes: 1) extracting accurate quantity take off data from BIM models, and 2) manipulating cost analysis results to provide informative feedback. Some recent efforts involve developing detailed definitions, enhanced IFC-based formats and in-house standards for assemblies that encompass building models (e.g. US Corps of Engineers). Some commercial applications enhance the level of detail associated to BIM objects with assembly descriptions to produce lightweight BIM models that can be used by different applications for various purposes (e.g. Autodesk for design review, Navisworks for scheduling, Innovaya for visual estimating, etc.). This study suggests the integration of design and analysis tools by means of managing all building data in one shared repository accessible to multiple domains in the AEC industry (Eastman, 1999; Eastman et al., 2008; authors, 2010). Our approach aims at providing an integrated platform that incorporates a quantity take off extraction method from IFC models, a cost analysis model, and a comprehensive cost reporting scheme, using the Solibri Model Checker (SMC) development environment. Approach As part of the effort to improve the performance of federal buildings, GSA evaluates concept design alternatives based on their compliance with specific requirements, including cost analysis. Two basic challenges emerge in the process of automating cost analysis for BIM models: 1) At this early concept design stage, only minimal information is available to produce a reliable analysis, such as space names and areas, and building gross area, 2) design alternatives share a lot of programmatic requirements such as location, functional spaces and other data. It is thus crucial to integrate other factors that contribute to substantial cost differences such as perimeter, and exterior wall and roof areas. These are extracted from BIM models using IFC data and input through XML into the Parametric Cost Engineering System (PACES, 2010) software to generate cost analysis reports. PACES uses this limited dataset at a conceptual stage and RSMeans (2010) data to infer cost assemblies at different levels of detail. Functionalities Cost model import module The cost model import module has three main functionalities: generating the input dataset necessary for the cost model, performing a semantic mapping between building type specific names and name aggregation structures in PACES known as functional space areas (FSAs), and managing cost data external to the BIM model, such as location and construction duration. The module computes building data such as footprint, gross area, perimeter, external wall and roof area and building space areas. This data is generated through SMC in the form of an XML file and imported into PACES. Reporting module The reporting module uses the cost report generated by PACES to develop a comprehensive report in the form of an excel spreadsheet. This report consists of a systems-elemental estimate that shows the main systems of the building in terms of UniFormat categories, escalation, markups, overhead and conditions, a UniFormat Level III report, and a cost breakdown that provides a summary of material, equipment, labor and total costs. Building parameters are integrated in the report to provide insight on the variations among design alternatives.
keywords building information modeling, interoperability, cost analysis, IFC
series CAAD Futures
email
last changed 2012/02/11 19:21

_id b4d2
authors Caldas, Luisa G. and Norford, Leslie K.
year 1999
title A Genetic Algorithm Tool for Design Optimization
doi https://doi.org/10.52842/conf.acadia.1999.260
source Media and Design Process [ACADIA ‘99 / ISBN 1-880250-08-X] Salt Lake City 29-31 October 1999, pp. 260-271
summary Much interest has been recently devoted to generative processes in design. Advances in computational tools for design applications, coupled with techniques from the field of artificial intelligence, have lead to new possibilities in the way computers can inform and actively interact with the design process. In this paper we use the concepts of generative and goal-oriented design to propose a computer tool that can help the designer to generate and evaluate certain aspects of a solution towards an optimized behavior of the final configuration. This work focuses mostly on those aspects related to the environmental performance of the building. Genetic Algorithms are applied as a generative and search procedure to look for optimized design solutions in terms of thermal and lighting performance in a building. The Genetic Algorithm (GA) is first used to generate possible design solutions, which are then evaluated in terms of lighting and thermal behavior using a detailed thermal analysis program (DOE2.1E). The results from the simulations are subsequently used to further guide the GA search towards finding low-energy solutions to the problem under study. Solutions can be visualized using an AutoLisp routine. The specific problem addressed in this study is the placing and sizing of windows in an office building. The same method is applicable to a wide range of design problems like the choice of construction materials, design of shading elements, or sizing of lighting and mechanical systems for buildings.
series ACADIA
email
last changed 2022/06/07 07:54

_id d9d0
authors Cohen Egler, Tamara Tania
year 1999
title Río Digital (Digital Rio)
source III Congreso Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings] Montevideo (Uruguay) September 29th - October 1st 1999, pp. 478-481
summary RioDigital is a text situated on new forms of expression of knowledge over the city. It is a written multimedia whose objective is to place in disposability for the society the complexity of urban space on its multiple historical determinations, of its space, social and cultural forms. It is a research over the potentials of digital art to express the processes of constitution of social forms and constructions of urban space. The motion of works was in the sense of using this language to reconstitute and vivify the history of Rio de Janeiro city in the XX century. The city is an ensemble of symbols that encounters the language in its best form of presentation. The research identified visual documents as films, photos and maps that made possible to reconstruct processes of transformation, worked through the use of digital images technology that allows expression and turns move perceptible the transmission of this history. The digital image is certainly a possibility to represent urban reality. Through movement, illumination of image and of writing it was possible to express to process of construction and reconstruction of space building and social changes. We understand that the condition of citizen is associated with the feeling of belonging, which urban process every time move complex and difficult to understand, that new technologies can through synthesis, connectivity and interactivity expand the capacity of indivils to know the city and act positively with it. It is an intention to amplify the sense of belonging and encourage the action of transform.
series SIGRADI
email
last changed 2016/03/10 09:49

_id 1b4d
authors Ding, Lan
year 1999
title An Evolutionary Model for Style Representation Emergence in Design
source University of Sydney, Key Centre of Design Computing and Cognition
summary This thesis is concerned with the development of an evolutionary process model for style representation emergence in design. It explores issues involved in the interpretation of style, the concept and process of style representation emergence, an evolutionary approach based on genetic engineering, and its computational implementation. Style is a complex phenomenon in design. Interpreting and formulating design style is a difficult task. This thesis proposes a language model which interprets style space utilising hierarchical levels that map onto syntax and semantics. The style space is then formulated using a genetic description. Current studies have discussed shape semantics emergence in design, but none has been proposed for the emergence of style representation. This thesis provides the concept of style representation emergence with the emphasis on the interpretative aspect of style as well as the emergence process. It explores the emergence process of style representation through an evolutionary approach. Simulation of biological evolution appears to be very useful for design problems. This thesis develops style representation emergence through evolutionary simulation based on genetic engineering. A hierarchical evolutionary process encompassing competition as well as discovery and an evolutionary combination is proposed and developed. A computational representation of style can then be derived by the computer system through the use of this evolutionary process. This model of style representation emergence is applied to traditional Chinese architecture. An evolutionary system is implemented and presented with some examples of traditional Chinese architectural facades. The results from the implementation of the system are analysed and the utility of this model is investigated. The implementation is developed in a Unix environment using the C language. The AutoCAD package is used for the graphic representation.

series thesis:PhD
email
last changed 2003/05/15 07:25

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 31HOMELOGIN (you are user _anon_876225 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002