CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 542

_id avocaad_2001_02
id avocaad_2001_02
authors Cheng-Yuan Lin, Yu-Tung Liu
year 2001
title A digital Procedure of Building Construction: A practical project
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In earlier times in which computers have not yet been developed well, there has been some researches regarding representation using conventional media (Gombrich, 1960; Arnheim, 1970). For ancient architects, the design process was described abstractly by text (Hewitt, 1985; Cable, 1983); the process evolved from unselfconscious to conscious ways (Alexander, 1964). Till the appearance of 2D drawings, these drawings could only express abstract visual thinking and visually conceptualized vocabulary (Goldschmidt, 1999). Then with the massive use of physical models in the Renaissance, the form and space of architecture was given better precision (Millon, 1994). Researches continued their attempts to identify the nature of different design tools (Eastman and Fereshe, 1994). Simon (1981) figured out that human increasingly relies on other specialists, computational agents, and materials referred to augment their cognitive abilities. This discourse was verified by recent research on conception of design and the expression using digital technologies (McCullough, 1996; Perez-Gomez and Pelletier, 1997). While other design tools did not change as much as representation (Panofsky, 1991; Koch, 1997), the involvement of computers in conventional architecture design arouses a new design thinking of digital architecture (Liu, 1996; Krawczyk, 1997; Murray, 1997; Wertheim, 1999). The notion of the link between ideas and media is emphasized throughout various fields, such as architectural education (Radford, 2000), Internet, and restoration of historical architecture (Potier et al., 2000). Information technology is also an important tool for civil engineering projects (Choi and Ibbs, 1989). Compared with conventional design media, computers avoid some errors in the process (Zaera, 1997). However, most of the application of computers to construction is restricted to simulations in building process (Halpin, 1990). It is worth studying how to employ computer technology meaningfully to bring significant changes to concept stage during the process of building construction (Madazo, 2000; Dave, 2000) and communication (Haymaker, 2000).In architectural design, concept design was achieved through drawings and models (Mitchell, 1997), while the working drawings and even shop drawings were brewed and communicated through drawings only. However, the most effective method of shaping building elements is to build models by computer (Madrazo, 1999). With the trend of 3D visualization (Johnson and Clayton, 1998) and the difference of designing between the physical environment and virtual environment (Maher et al. 2000), we intend to study the possibilities of using digital models, in addition to drawings, as a critical media in the conceptual stage of building construction process in the near future (just as the critical role that physical models played in early design process in the Renaissance). This research is combined with two practical building projects, following the progress of construction by using digital models and animations to simulate the structural layouts of the projects. We also tried to solve the complicated and even conflicting problems in the detail and piping design process through an easily accessible and precise interface. An attempt was made to delineate the hierarchy of the elements in a single structural and constructional system, and the corresponding relations among the systems. Since building construction is often complicated and even conflicting, precision needed to complete the projects can not be based merely on 2D drawings with some imagination. The purpose of this paper is to describe all the related elements according to precision and correctness, to discuss every possibility of different thinking in design of electric-mechanical engineering, to receive feedback from the construction projects in the real world, and to compare the digital models with conventional drawings.Through the application of this research, the subtle relations between the conventional drawings and digital models can be used in the area of building construction. Moreover, a theoretical model and standard process is proposed by using conventional drawings, digital models and physical buildings. By introducing the intervention of digital media in design process of working drawings and shop drawings, there is an opportune chance to use the digital media as a prominent design tool. This study extends the use of digital model and animation from design process to construction process. However, the entire construction process involves various details and exceptions, which are not discussed in this paper. These limitations should be explored in future studies.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id e6ca
authors Curry, Michael
year 1998
title Digital Places: Living With Geographic Information Technologies
source N.Y.: Routledge
summary The last twenty-five years have seen major changes in the nature and scope of geographical information. This has happened in one way in society at large, where computers, satellites and global positioning systems have made geographical information more extensive, more detailed and more available. It has happened in another way within the university, where rapidly evolving geographic information systems have been touted as tools useful in a wide range of disciplines, tools that will resolve problems as different as the nature of global climate change and the routing of mail. In both settings the move from manual to computer-based systems is viewed as having a natural trajectory, from less powerful to more powerful technologies. These systems are held to be increasingly able to model and represent all that is important in geographical knowledge and behaviour. They are seen as fitting into and supporting traditional scientific and social practices and institutions. Digital Places: Living with Geographic Information Technologies shows that on each score the systems have been misunderstood and their impacts underestimated. By offering an understanding of Geographic Information Systems within the social, economic, legal, political and ethical contexts within which they exist, the author shows that there are substantial limits to their ability to represent the very objects and relationships, people and places, that many believe to be most important. Focusing on the ramifications of GIS usage, Digital Places shows that they are associated with far-reaching changes in the institutions in which they exist, and in the lives of those they touch. In the end they call for a complete rethinking of basic ideas, like privacy and intellectual property and the nature of scientific practice, that have underpinned public life for the last one hundred years.
series other
last changed 2003/04/23 15:14

_id ga0021
id ga0021
authors Eacott, John
year 2000
title Generative music composition in practice - a critical evaluation
source International Conference on Generative Art
summary This critical evaluation will discuss 4 computer based musical works which, for reasons I shall explain, I describe as non-linear or generative. The works have been constructed by me and publicly performed or exhibited during a two year period from October 1998 to October 2000. ‘In the beginning…’ interactive music installation, strangeAttraction, Morley Gallery, London. July 1999 ‘jnrtv’ live generative dance music May 1999 to Dec 2000 ‘jazz’ interactive music installation, another strangeAttraction Morley Gallery, London. July 2000-09-26 ‘the street’ architectural interactive music installation, University of Westminster Oct 2000 Introduction I have always loved the practice of composing, particularly when it means scoring a work to be played by a live ensemble. There is something about taking a fresh sheet of manuscript , ruling the bar lines, adding clefs, key and time signatures and beginning the gradual process of adding notes, one at a time to the score until it is complete that is gratifying and compensates for the enormous effort involved. The process of scoring however is actually one distinct act within the more general task of creating music. Recently, the notion of ‘composing’ has met challenges through an increased interest in non-linear compositional methods. It is actually the presence of Chaotic or uncontrolable elements which add real beauty to music and many if not all of the things we value. If we think of a sunset, waves lapping on the shore, plants, trees a human face and the sound of the human voice, these things are not perfect and more importantly perhaps, they are transient, constantly changing and evolving. Last year and again this year, I have organised an exhibition of interactive , non-linear music installations called 'strangeAttraction'. The title refers to what Edward Lorenz called a ‘strange attractor’ the phenomenon that despite vast degrees of Chaos and uncertainty within a system, there is a degree of predictability, the tendency for chaotic behaviour to ‘attract’ towards a probable set of outcomes. Composition that deals with 'attractors' or probable outcomes rather than specific details which are set in stone is an increasingly intriguing notion.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 2587
authors Gong, Yihong
year 1998
title Intelligent image databases
source Boston, Kluwer
summary Intelligent Image Databases: Towards Advanced Image Retrieval addresses the image feature selection issue in developing content-based image retrieval systems. The book first discusses the four important issues in developing a complete content-based image retrieval system, and then demonstrates that image feature selection has significant impact on the remaining issues of system design. Next, it presents an in-depth literature survey on typical image features explored by contemporary content-based image retrieval systems for image matching and retrieval purposes. The goal of the survey is to determine the characteristics and the effectiveness of individual features, so as to establish guidelines for future development of content-based image retrieval systems. Intelligent Image Databases: Towards Advanced Image Retrieval describes the Advanced Region-Based Image Retrieval System (ARBIRS) developed by the authors for color images of real-world scenes. They have selected image regions for building ARBIRS as the literature survey suggests that prominent image regions, along with their associated features, provide a higher probability for achieving a higher level content-based image retrieval system. A major challenge in building a region-based image retrieval system is that prominent regions are rather difficult to capture in an accurate and error-free condition, particularly those in images of real-world scenes. To meet this challenge, the book proposes an integrated approach to tackle the problem via feature capturing, feature indexing, and database query. Through comprehensive system evaluation, it is demonstrated how these systematically integrated efforts work effectively to accomplish advanced image retrieval. Intelligent Image Databases: Towards Advanced Image Retrieval serves as an excellent reference and may be used as a text for advanced courses on the topic.
series other
last changed 2003/04/23 15:14

_id ee96
authors Johnson, Scott
year 1998
title Making Models Architectural: Protean Representations to Fit Architects’ Minds
doi https://doi.org/10.52842/conf.acadia.1998.354
source Digital Design Studios: Do Computers Make a Difference? [ACADIA Conference Proceedings / ISBN 1-880250-07-1] Québec City (Canada) October 22-25, 1998, pp. 354-365
summary A rich vocabulary has evolved for describing architecture. It serves not only as a means of communication, but also as an embodiment of concepts relating to form, space, structure, function, mood, and symbolism. We architects not only speak in terms of walls, rooms, roofs, arches, etc., we see in terms of them and think in terms of them, as well. Such concepts are integral to our ability to design. Typical CAD representations, however, are based on geometric/mathematical elements like points, lines, planes, and symbols. Even more experimental approaches like parametric shapes or procedural assemblies correspond poorly to architectural elements, and seldom lend themselves well to making conceptual changes that would allow exploration of design alternatives. Small wonder some architecture schools experience a division between computer and studio courses, or even between computer and studio faculty. Different ways of talking and thinking are involved. The concepts involved are often mutually exclusive. This paper discusses an attempt to address this conceptual mismatch, using what are termed “protean” (meaning “very changeable”) elements. These are high-level elements corresponding to architectural concepts like “wall,” or “dome.” They each have parameters appropriate for the particular type of element they represent, and produce the polyhedra necessary for graphics based on these parameters. A system is being implemented to allow models to be constructed using these elements. The protean elements form a loosely structured model, in which some elements hierarchically contain others, and some elements are essentially freestanding, being created and manipulated independently of other elements. Characteristics of protean element are discussed, including the underlying object-oriented structure, the relationship between elements and graphics, and functions associated with the objects. A scheme is explained whereby all parts of a design can be represented even when the design includes extremely unusual forms not conforming to predictable classes of elements. The necessary support framework is also discussed; general flow of the system and mechanisms for viewing the model and editing subcomponents are explained. The current status of the project, and intentions for future work are discussed. The project has been partially implemented, and the necessary framework to support the system is mostly complete.

series ACADIA
email
last changed 2022/06/07 07:52

_id 68fb
authors Khemlani, L., Timerman, A., Benne, B. and Kalay, Y.E.
year 1998
title Intelligent representation for computer-aided building design
source Automation in Construction 8 (1) (1998) pp. 49-71
summary At the core of any computational system that can support design development, analysis, and evaluation is an “intelligent” building representation which should be able to represent all the different components that make up a building, along with the manner in which they come together. In other words, the representation must be informationally complete and semantically rich. The paper discusses these two criteria and briefly reviews other research efforts aimed at developing building representations for computer-aided design that attempt to meet them. Our solution to this problem is then presented. It is aimed primarily at the schematic design phase, the rationale for which is also stated. Taking the view that buildings are unique assemblies of discrete, mostly standardized components, our representation is clearly divided into two components: the Object Database (ODB) which stores detailed information about various building elements, and the Project Database (PDB) which holds information about how these elements are assembled to make up a particular building. An ODB may be shared by many building projects, while the PDB must necessarily be unique to each. The data schemas of both the PDB and the ODB are described in detail and their computational implementation, to the extent that it has been completed, is illustrated.
series journal paper
last changed 2003/03/05 13:12

_id ddss9834
id ddss9834
authors Kovács, László Béla and Kotsis, István
year 1998
title Basic Concepts and Prototypes of a Land Usage Design and Decision Support System
source Timmermans, Harry (Ed.), Fourth Design and Decision Support Systems in Architecture and Urban Planning Maastricht, the Netherlands), ISBN 90-6814-081-7, July 26-29, 1998
summary This paper presents the basic ideas of a computer system for supporting urban design and decisions on land use. We argue, that the high complexity of urban design - inherent in the its large number of interdependent views and aspects - seems to justify a flexible support system capable of reasoning and conceptual modelling. Such a system may be prohibitively resource demanding unless we are able to build it up from smaller and larger modules of different types and functionality and which canbe created basically in an incremental way without a complete plan in advance. Two prototypes concerning urban designs and a small flexible design rule interpreter/handler is presented for free standing buildings.
series DDSS
email
last changed 2003/08/07 16:36

_id c38b
authors Kunz, J.C., Christiansen, T.R., Cohen, G.P., Jin, Y. and Levitt, R.E.
year 1998
title The Virtual Design Team
source Communications of The ACM, Vol. 41, No. 11, November, 1998
summary The long range goal of the Virtual Design Team" (VDT) research program is to develop computational tools to analyze decision making and communication behavior and thereby to support true organizational (re)engineering. This article introduces the underlying theory, the implementation of the theory as a computational model, and results from industrial test cases. Organization theory traditionally describes organizations only at an aggregate-level, describing and predicting the behavior of entire organizations in terms of general qualitative predictions. We define and implement a "micro" theory of the structure and behavior of components of organizations, explicitly representing activities, groups of people called "actors," and organizational structure and policies for project teams. A VDT model can be "run" by a discrete event simulation. Emergent aggregate model output behaviors include the predicted time to complete a project, the total effort to do the project, and a measure of process quality. More detailed model behaviors include the time-varying backlog of individual actors and the "exceptions" associated with activities. The results are detailed and specific, so they can guide specific managerial interventions in a project team and can support sensitivity studies of the relative impact of different organizational changes. We conclude that such a theory is tractable and predictive for complex but relatively routine, project-oriented design tasks. The application for which VDT offers unique new kinds of insights is where an organization is striving to shrink time to market dramatically for a product that is similar to ones it has previously developed. Reducing time to market dramatically almost always requires that previously sequential activities are executed more concurrently. In this situation, experienced managers can still correctly identify the required activities and estimate their durations and skill requirements; but they almost always underestimate the increased workload arising from exponentially higher coordination needs and the propagation of rework between the now highly concurrent activities. The VDT framework, which explicitly models information dependency and failure propagation between concurrent activities, has proven to be far more accurate, and to incorporate a wider range of parameters, than CPM/PERT process models for these fast-paced development projects."
series journal paper
last changed 2003/04/23 15:50

_id a616
authors Mallot - Hanspeter, A. and Gillner, S.
year 1998
title Navigation and acquisition of spatial knowledge in a virtual maze
source Journal of Cognitive Neuro-Science, 10, pp. 445-463
summary Spatial behavior in humans and animals includes a wide variety of behavioral competences and makes use of a large number of sensory cues. Here we studied the ability of human subjects to search locations, to find shortcuts and novel paths, to estimate distances between remembered places, and to draw sketch maps of the explored environment; these competences are related to goal-independent memory of space, or cognitive maps. Information on spatial relations was restricted to two types: a visual motion sequence generated by simulated movements in a virtual maze and the subject's own movement decisions defining the path through the maze. Visual information was local (i.e., no global landmarks or compass information was provided). Other position and movement information (vestibular or proprioceptive) was excluded. The amount of visual information provided was varied over four experimental conditions. The results indicate that human subjects are able to learn a virtual maze from sequences of local views and movements. The information acquired is local, consisting of recognized positions and movement decisions associated to them. Although simple associations of this type can be shown to be present in some subjects, more complete configurational knowledge is acquired as well. The results are discussed in a view-based framework of navigation and the representation of spatial knowledge by means of a view graph.
series journal paper
last changed 2003/04/23 15:50

_id 37
authors Morelli, RubÈn DarÌo and Marina, Cristian
year 1998
title Geometria y Grafica Digital Como Reflexion y Racionalizacion Del Proyecto Arquitectonico (Geometry and Digital Graphics as Reflexion and Rationalization of the Architectural Project)
source II Seminario Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings / ISBN 978-97190-0-X] Mar del Plata (Argentina) 9-11 september 1998, pp. 280-285
summary The methodology of the work consists in the following: (a) Starting from the photographic image of an architectonic work (Santisimo Sacramento Church situated in 3451 Bv. OroÒo street, Rosario city, Santa Fe, Republica Argentina), and applying the rules of Descriptive Geometry. about photographic perspective, rebuild - restore the orthogonal parallel projection of its facade. (b) Once the restitution is done, introduce the information into the computer, in order to: Make a geometric analysis of the architectonic shape, applying 2D systems (bidimensional diedric or Monge method ) and 3D (tridimensional, wireframes and renders); Obtain a complete 3D image of the Tower, that means the virtual model of the real object.
series SIGRADI
email
last changed 2016/03/10 09:55

_id 00eb
authors Morozumi, Mitsuo
year 1998
title Gradual Introduction of CAAD to Develop and Support Students’ Ability in Design Studio
doi https://doi.org/10.52842/conf.ecaade.1998.107
source Computers in Design Studio Teaching [EAAE/eCAADe International Workshop Proceedings / ISBN 09523687-7-3] Leuven (Belgium) 13-14 November 1998, pp. 107-114
summary How to integrate CAAD into design education, and how to teach CAAD as a tool of design thinking has been a difficult issue left unresolved for architecture schools. This paper discusses the possible approach to these issues which were experienced in a Japanese university. In the first section, it will summarize the present situation of CAAD education in Japan. The second section reports the framework of design education and roles of CAAD programmed in a department. The third section introduces an example of course programs and students’ work. The fourth section observes a recent outcome of experimented program, such as, students’ work, scores and some results of a student questionnaire. In the final section, it will discuss the approach that introduce different levels of CAAD usage to design studios as the level of design education advances, was successful to integrate CAAD into design education.
series eCAADe
email
more http://www.eaae.be/
last changed 2022/06/07 07:59

_id 06cb
authors Pfleiderer, Glenn M.
year 1998
title The Softdesk Primer
source Prentice Hall
summary This tutorial guides readers through a complete project--from beginning to final project output. It develops a solid understanding of program and project flow and a mastery of Softdesk module commands.
series other
last changed 2003/02/26 18:58

_id 48db
authors Proctor, George
year 2001
title CADD Curriculum - The Issue of Visual Acuity
doi https://doi.org/10.52842/conf.ecaade.2001.192
source Architectural Information Management [19th eCAADe Conference Proceedings / ISBN 0-9523687-8-1] Helsinki (Finland) 29-31 August 2001, pp. 192-200
summary Design educators attempt to train the eyes and minds of students to see and comprehend the world around them with the intention of preparing those students to become good designers, critical thinkers and ultimately responsible architects. Over the last eight years we have been developing the digital media curriculum of our architecture program with these fundamental values. We have built digital media use and instruction on the foundation of our program which has historically been based in physical model making. Digital modeling has gradually replaced the capacity of physical models as an analytical and thinking tool, and as a communication and presentation device. The first year of our program provides a foundation and introduction to 2d and 3d design and composition, the second year explores larger buildings and history, the third year explores building systems and structure through design studies of public buildings, fourth year explores urbanism, theory and technology through topic studios and, during the fifth year students complete a capstone project. Digital media and CADD have and are being synchronized with the existing NAAB accredited regimen while also allowing for alternative career options for students. Given our location in the Los Angeles region, many students with a strong background in digital media have gone on to jobs in video game design and the movie industry. Clearly there is much a student of architecture must learn to attain a level of professional competency. A capacity to think visually is one of those skills and is arguably a skill that distinguishes members of the visual arts (including Architecture) from other disciplines. From a web search of information posted by the American Academy of Opthamology, Visual Acuity is defined as an ability to discriminate fine details when looking at something and is often measured with the Snellen Eye Chart (the 20/20 eye test). In the context of this paper visual acuity refers to a subject’s capacity to discriminate useful abstractions in a visual field for the purposes of Visual Thinking- problem solving through seeing (Arnheim, 1969, Laseau 1980, Hoffman 1998). The growing use of digital media and the expanding ability to assemble design ideas and images through point-and-click methods makes the cultivation and development of visual skills all the more important to today’s crop of young architects. The advent of digital media also brings into question the traditional, static 2d methods used to build visual skills in a design education instead of promoting active 3d methods for teaching, learning and developing visual skills. Interactive digital movies provide an excellent platform for promoting visual acuity, and correlating the innate mechanisms of visual perception with the abstractions and notational systems used in professional discourse. In the context of this paper, pedagogy for building visual acuity is being considered with regard to perception of the real world, for example the visual survey of an environment, a site or a street scene and how that visual survey works in conjunction with practice.
keywords Curriculum, Seeing, Abstracting, Notation
series eCAADe
email
last changed 2022/06/07 08:00

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 497c
authors Savage, J., Miles, C., Moore, C. and Miles, J.
year 1998
title The interaction of time and cost constraints on the design process
source Design Studies, 19[2]: 217-233, Apr
summary This paper reports an experimental study investigating the influence of both `external' and `task inherent' constraints on simple design tasks. Subjects were required to complete each of 3 tasks in one of 4 conditions. The conditions comprised a control condition (no time or cost constraints imposed), a time constrained condition, a cost constrained condition and a time and cost constrained condition. The tasks required subjects to build structures predominantly from sheets of A4. For example, task one required subjects to `create a self supporting structure, which has a minimum height of 19 cm.' The pattern of results showed that the constraints of time and cost affected the design process both independently and in combination. In addition, the range of designs, completion times, and design costs were all affected differentially.
series journal paper
last changed 2003/04/23 15:50

_id avocaad_2001_19
id avocaad_2001_19
authors Shen-Kai Tang, Yu-Tung Liu, Yu-Sheng Chung, Chi-Seng Chung
year 2001
title The visual harmony between new and old materials in the restoration of historical architecture: A study of computer simulation
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In the research of historical architecture restoration, scholars respectively focus on the field of architectural context and architectural archeology (Shi, 1988, 1990, 1991, 1992, 1995; Fu, 1995, 1997; Chiu, 2000) or on architecture construction and the procedure of restoration (Shi, 1988, 1989; Chiu, 1990). How to choose materials and cope with their durability becomes an important issue in the restoration of historical architecture (Dasser, 1990; Wang, 1998).In the related research of the usage and durability of materials, some scholars deem that, instead of continuing the traditional ways that last for hundreds of years (that is to replace new materials with old ones), it might be better to keep the original materials (Dasser, 1990). However, unavoidably, some of the originals are much worn. Thus we have to first establish the standard of eliminating components, and secondly to replace identical or similar materials with the old components (Lee, 1990). After accomplishing the restoration, we often unexpectedly find out that the renewed historical building is too new that the sense of history is eliminated (Dasser, 1990; Fu, 1997). Actually this is the important factor that determines the accomplishment of restoration. In the past, some scholars find out that the contrast and conflict between new and old materials are contributed to the different time of manufacture and different coating, such as antiseptic, pattern, etc., which result in the discrepancy of the sense of visual perception (Lee, 1990; Fu, 1997; Dasser, 1990).In recent years, a number of researches and practice of computer technology have been done in the field of architectural design. We are able to proceed design communication more exactly by the application of some systematic softwares, such as image processing, computer graphic, computer modeling/rendering, animation, multimedia, virtual reality and so on (Lawson, 1995; Liu, 1996). The application of computer technology to the research of the preservation of historical architecture is comparatively late. Continually some researchers explore the procedure of restoration by computer simulation technology (Potier, 2000), or establish digital database of the investigation of historical architecture (Sasada, 2000; Wang, 1998). How to choose materials by the technology of computer simulation influences the sense of visual perception. Liu (2000) has a more complete result on visual impact analysis and assessment (VIAA) about the research of urban design projection. The main subjects of this research paper focuses on whether the technology of computer simulation can extenuate the conflict between new and old materials that imposed on visual perception.The objective of this paper is to propose a standard method of visual harmony effects for materials in historical architecture (taking the Gigi Train Station destroyed by the earthquake in last September as the operating example).There are five steps in this research: 1.Categorize the materials of historical architecture and establish the information in digital database. 2.Get new materials of historical architecture and establish the information in digital database. 3.According to the mixing amount of new and old materials, determinate their proportion of the building; mixing new and old materials in a certain way. 4.Assign the mixed materials to the computer model and proceed the simulation of lighting. 5.Make experts and the citizens to evaluate the accomplished computer model in order to propose the expected standard method.According to the experiment mentioned above, we first address a procedure of material simulation of the historical architecture restoration and then offer some suggestions of how to mix new and old materials.By this procedure of simulation, we offer a better view to control the restoration of historical architecture. And, the discrepancy and discordance by new and old materials can be released. Moreover, we thus avoid to reconstructing ¡§too new¡¨ historical architecture.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id avocaad_2001_20
id avocaad_2001_20
authors Shen-Kai Tang
year 2001
title Toward a procedure of computer simulation in the restoration of historical architecture
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In the field of architectural design, “visualization¨ generally refers to some media, communicating and representing the idea of designers, such as ordinary drafts, maps, perspectives, photos and physical models, etc. (Rahman, 1992; Susan, 2000). The main reason why we adopt visualization is that it enables us to understand clearly and to control complicated procedures (Gombrich, 1990). Secondly, the way we get design knowledge is more from the published visualized images and less from personal experiences (Evans, 1989). Thus the importance of the representation of visualization is manifested.Due to the developments of computer technology in recent years, various computer aided design system are invented and used in a great amount, such as image processing, computer graphic, computer modeling/rendering, animation, multimedia, virtual reality and collaboration, etc. (Lawson, 1995; Liu, 1996). The conventional media are greatly replaced by computer media, and the visualization is further brought into the computerized stage. The procedure of visual impact analysis and assessment (VIAA), addressed by Rahman (1992), is renewed and amended for the intervention of computer (Liu, 2000). Based on the procedures above, a great amount of applied researches are proceeded. Therefore it is evident that the computer visualization is helpful to the discussion and evaluation during the design process (Hall, 1988, 1990, 1992, 1995, 1996, 1997, 1998; Liu, 1997; Sasada, 1986, 1988, 1990, 1993, 1997, 1998). In addition to the process of architectural design, the computer visualization is also applied to the subject of construction, which is repeatedly amended and corrected by the images of computer simulation (Liu, 2000). Potier (2000) probes into the contextual research and restoration of historical architecture by the technology of computer simulation before the practical restoration is constructed. In this way he established a communicative mode among archeologists, architects via computer media.In the research of restoration and preservation of historical architecture in Taiwan, many scholars have been devoted into the studies of historical contextual criticism (Shi, 1988, 1990, 1991, 1992, 1995; Fu, 1995, 1997; Chiu, 2000). Clues that accompany the historical contextual criticism (such as oral information, writings, photographs, pictures, etc.) help to explore the construction and the procedure of restoration (Hung, 1995), and serve as an aid to the studies of the usage and durability of the materials in the restoration of historical architecture (Dasser, 1990; Wang, 1998). Many clues are lost, because historical architecture is often age-old (Hung, 1995). Under the circumstance, restoration of historical architecture can only be proceeded by restricted pictures, written data and oral information (Shi, 1989). Therefore, computer simulation is employed by scholars to simulate the condition of historical architecture with restricted information after restoration (Potier, 2000). Yet this is only the early stage of computer-aid restoration. The focus of the paper aims at exploring that whether visual simulation of computer can help to investigate the practice of restoration and the estimation and evaluation after restoration.By exploring the restoration of historical architecture (taking the Gigi Train Station destroyed by the earthquake in last September as the operating example), this study aims to establish a complete work on computer visualization, including the concept of restoration, the practice of restoration, and the estimation and evaluation of restoration.This research is to simulate the process of restoration by computer simulation based on visualized media (restricted pictures, restricted written data and restricted oral information) and the specialized experience of historical architects (Potier, 2000). During the process of practicing, communicates with craftsmen repeatedly with some simulated alternatives, and makes the result as the foundation of evaluating and adjusting the simulating process and outcome. In this way we address a suitable and complete process of computer visualization for historical architecture.The significance of this paper is that we are able to control every detail more exactly, and then prevent possible problems during the process of restoration of historical architecture.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id ddss9860
id ddss9860
authors Vakalo, E-G. and Fahmy, A.
year 1998
title A Theoretical Framework for the Analysis and Derivation of Orthogonal Building Plans and Sections
source Timmermans, Harry (Ed.), Fourth Design and Decision Support Systems in Architecture and Urban Planning Maastricht, the Netherlands), ISBN 90-6814-081-7, July 26-29, 1998
summary Architects are generally perceived as “Formgivers with an extraordinary gift” (Ackerman, 1980:12). Implicit in this statement is the belief that the operations that architects employ to compose their designs are the product of a creative faculty that is beyond the reach of rational discourse, and thereby cannot be subjected to logical investigation. This view is detrimental to the advancement of knowledge about architectural composition and adversely affects both practice and education in architecture. More specifically, it prevents the architectural community from acquiring of a more refined conception about how architects derive their designs. In contrast to this view, this study demonstrates that architectural form-making is amenable to logical analysis. In specific, this is to be done through a theoretical and computational framework that describe and explain the tasks involved in the making of orthogonal building plans and sections. In addition to illustrating the susceptibility of architectural form-making to logical analysis, the frameworks proposed in this study overcome the limitations of previously established theories thatdeal with architectural form-making. These can be divided into two categories: normative and positive theories.Normative theories include architectural treatises and manifestos. A major limitation of normativetheories is that they have limited explanatory power. Their concern is with promoting a specific aesthetic ideology and prescribing rules that can be used to derive compositions that conform to it. Therefore, they cannot be used to explain form-making in general. Positive frameworks, such asshape grammar, rely on rules to describe derivation and analysis processes. Nevertheless, they do not provide a comprehensive description of the tasks involved in architectural form-making. This causes the relation between the rules and compositional tasks to be ambiguous. It also affects adversely the ability of these frameworks to provide architects with a complete understanding of the role of compositional rules in derivation or analysis processes.
series DDSS
type normal paper
last changed 2010/05/16 09:11

_id e336
authors Achten, H., Roelen, W., Boekholt, J.-Th., Turksma, A. and Jessurun, J.
year 1999
title Virtual Reality in the Design Studio: The Eindhoven Perspective
doi https://doi.org/10.52842/conf.ecaade.1999.169
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 169-177
summary Since 1991 Virtual Reality has been used in student projects in the Building Information Technology group. It started as an experimental tool to assess the impact of VR technology in design, using the environment of the associated Calibre Institute. The technology was further developed in Calibre to become an important presentation tool for assessing design variants and final design solutions. However, it was only sporadically used in student projects. A major shift occurred in 1997 with a number of student projects in which various computer technologies including VR were used in the whole of the design process. In 1998, the new Design Systems group started a design studio with the explicit aim to integrate VR in the whole design process. The teaching effort was combined with the research program that investigates VR as a design support environment. This has lead to increasing number of innovative student projects. The paper describes the context and history of VR in Eindhoven and presents the current set-UP of the studio. It discusses the impact of the technology on the design process and outlines pedagogical issues in the studio work.
keywords Virtual Reality, Design Studio, Student Projects
series eCAADe
email
last changed 2022/06/07 07:54

_id ddss9801
id ddss9801
authors Achten, Henri and Leeuwen, Jos van
year 1998
title A Feature-Based Description Technique for Design Processes: A Case Study
source Timmermans, Harry (Ed.), Fourth Design and Decision Support Systems in Architecture and Urban Planning Maastricht, the Netherlands), ISBN 90-6814-081-7, July 26-29, 1998
summary In order to develop appropriate tools for decision support in design processes, it is necessary to found them on an understanding of design. Analytical techniques of design processes that have a direct relationship with tool development can enhance design support systems development. The paper focuses on a design support system in the VR-DIS research program. The aim of this research program is to develop insight in the architectural design process and to establish design tools for architectsworking in Virtual Reality. The basic approach for data modelling in VR in this research is based on an extension of the Feature Based Modelling paradigm taken from design in mechanical engineering. The computer model of the design in the system is a Feature-based model. This paper describes design processes in terms of changes in the Feature-based model of the design. For this purpose, a case of a house design is used. Drawings in the conceptual design phase up to the preliminary design phase arestudied. Each state of the drawings is described in terms of a Feature-model. Particular design actions such as creation of spaces, definition of architectural elements, and changes during the design process can be expressed in terms of changes in the Feature-model. Because of the use of Features, the changes can be formalised in the VR-DIS system. The description in terms of Features offers an analytical toolthat leads to a functional brief for design support tools. The paper ends with a discussion of implications and future work.
series DDSS
last changed 2003/11/21 15:15

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 27HOMELOGIN (you are user _anon_775264 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002