CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 76

_id e234
authors Kalay, Yehuda E. and Harfmann, Anton C.
year 1985
title An Integrative Approach to Computer-Aided Design Education in Architecture
source February, 1985. [17] p. : [8] p. of ill
summary With the advent of CAD, schools of architecture are now obliged to prepare their graduates for using the emerging new design tools and methods in architectural practices of the future. In addition to this educational obligation, schools of architecture (possibly in partnership with practicing firms) are also the most appropriate agents for pursuing research in CAD that will lead to the development of better CAD software for use by the profession as a whole. To meet these two rather different obligations, two kinds of CAD education curricula are required: one which prepares tool- users, and another that prepares tool-builders. The first educates students about the use of CAD tools for the design of buildings, whereas the second educates them about the design of CAD tools themselves. The School of Architecture and Planning in SUNY at Buffalo has recognized these two obligations, and in Fall 1982 began to meet them by planning and implementing an integrated CAD environment. This environment now consists of 3 components: a tool-building sequence of courses, an advanced research program, and a general tool-users architectural curriculum. Students in the tool-building course sequence learn the principles of CAD and may, upon graduation, become researchers and the managers of CAD systems in practicing offices. While in school they form a pool of research assistants who may be employed in the research component of the CAD environment, thereby facilitating the design and development of advanced CAD tools. The research component, through its various projects, develops and provides state of the art tools to be used by practitioners as well as by students in the school, in such courses as architectural studio, environmental controls, performance programming, and basic design courses. Students in these courses who use the tools developed by the research group constitute the tool-users component of the CAD environment. While they are being educated in the methods they will be using throughout their professional careers, they also act as a 'real-world' laboratory for testing the software and thereby provide feedback to the research component. The School of Architecture and Planning in SUNY at Buffalo has been the first school to incorporate such a comprehensive CAD environment in its curriculum, thereby successfully fulfilling its obligation to train students in the innovative methods of design that will be used in architectural practices of the future, and at the same time making a significant contribution to the profession of architecture as a whole. This paper describes the methodology and illustrates the history of the CAD environment's implementation in the School
keywords CAD, architecture, education
series CADline
email
last changed 2003/06/02 13:58

_id c7e9
authors Maver, T.W.
year 2002
title Predicting the Past, Remembering the Future
source SIGraDi 2002 - [Proceedings of the 6th Iberoamerican Congress of Digital Graphics] Caracas (Venezuela) 27-29 november 2002, pp. 2-3
summary Charlas Magistrales 2There never has been such an exciting moment in time in the extraordinary 30 year history of our subject area, as NOW,when the philosophical theoretical and practical issues of virtuality are taking centre stage.The PastThere have, of course, been other defining moments during these exciting 30 years:• the first algorithms for generating building layouts (circa 1965).• the first use of Computer graphics for building appraisal (circa 1966).• the first integrated package for building performance appraisal (circa 1972).• the first computer generated perspective drawings (circa 1973).• the first robust drafting systems (circa 1975).• the first dynamic energy models (circa 1982).• the first photorealistic colour imaging (circa 1986).• the first animations (circa 1988)• the first multimedia systems (circa 1995), and• the first convincing demonstrations of virtual reality (circa 1996).Whereas the CAAD community has been hugely inventive in the development of ICT applications to building design, it hasbeen woefully remiss in its attempts to evaluate the contribution of those developments to the quality of the built environmentor to the efficiency of the design process. In the absence of any real evidence, one can only conjecture regarding the realbenefits which fall, it is suggested, under the following headings:• Verisimilitude: The extraordinary quality of still and animated images of the formal qualities of the interiors and exteriorsof individual buildings and of whole neighborhoods must surely give great comfort to practitioners and their clients thatwhat is intended, formally, is what will be delivered, i.e. WYSIWYG - what you see is what you get.• Sustainability: The power of «first-principle» models of the dynamic energetic behaviour of buildings in response tochanging diurnal and seasonal conditions has the potential to save millions of dollars and dramatically to reduce thedamaging environmental pollution created by badly designed and managed buildings.• Productivity: CAD is now a multi-billion dollar business which offers design decision support systems which operate,effectively, across continents, time-zones, professions and companies.• Communication: Multi-media technology - cheap to deliver but high in value - is changing the way in which we canexplain and understand the past and, envisage and anticipate the future; virtual past and virtual future!MacromyopiaThe late John Lansdown offered the view, in his wonderfully prophetic way, that ...”the future will be just like the past, onlymore so...”So what can we expect the extraordinary trajectory of our subject area to be?To have any chance of being accurate we have to have an understanding of the phenomenon of macromyopia: thephenomenon exhibitted by society of greatly exaggerating the immediate short-term impact of new technologies (particularlythe information technologies) but, more importantly, seriously underestimating their sustained long-term impacts - socially,economically and intellectually . Examples of flawed predictions regarding the the future application of information technologiesinclude:• The British Government in 1880 declined to support the idea of a national telephonic system, backed by the argumentthat there were sufficient small boys in the countryside to run with messages.• Alexander Bell was modest enough to say that: «I am not boasting or exaggerating but I believe, one day, there will bea telephone in every American city».• Tom Watson, in 1943 said: «I think there is a world market for about 5 computers».• In 1977, Ken Olssop of Digital said: «There is no reason for any individuals to have a computer in their home».The FutureJust as the ascent of woman/man-kind can be attributed to her/his capacity to discover amplifiers of the modest humancapability, so we shall discover how best to exploit our most important amplifier - that of the intellect. The more we know themore we can figure; the more we can figure the more we understand; the more we understand the more we can appraise;the more we can appraise the more we can decide; the more we can decide the more we can act; the more we can act themore we can shape; and the more we can shape, the better the chance that we can leave for future generations a trulysustainable built environment which is fit-for-purpose, cost-beneficial, environmentally friendly and culturally significactCentral to this aspiration will be our understanding of the relationship between real and virtual worlds and how to moveeffortlessly between them. We need to be able to design, from within the virtual world, environments which may be real ormay remain virtual or, perhaps, be part real and part virtual.What is certain is that the next 30 years will be every bit as exciting and challenging as the first 30 years.
series SIGRADI
email
last changed 2016/03/10 09:55

_id avocaad_2001_16
id avocaad_2001_16
authors Yu-Ying Chang, Yu-Tung Liu, Chien-Hui Wong
year 2001
title Some Phenomena of Spatial Characteristics of Cyberspace
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary "Space," which has long been an important concept in architecture (Bloomer & Moore, 1977; Mitchell, 1995, 1999), has attracted interest of researchers from various academic disciplines in recent years (Agnew, 1993; Benko & Strohmayer, 1996; Chang, 1999; Foucault, 1982; Gould, 1998). Researchers from disciplines such as anthropology, geography, sociology, philosophy, and linguistics regard it as the basis of the discussion of various theories in social sciences and humanities (Chen, 1999). On the other hand, since the invention of Internet, Internet users have been experiencing a new and magic "world." According to the definitions in traditional architecture theories, "space" is generated whenever people define a finite void by some physical elements (Zevi, 1985). However, although Internet is a virtual, immense, invisible and intangible world, navigating in it, we can still sense the very presence of ourselves and others in a wonderland. This sense could be testified by our naming of Internet as Cyberspace -- an exotic kind of space. Therefore, as people nowadays rely more and more on the Internet in their daily life, and as more and more architectural scholars and designers begin to invest their efforts in the design of virtual places online (e.g., Maher, 1999; Li & Maher, 2000), we cannot help but ask whether there are indeed sensible spaces in Internet. And if yes, these spaces exist in terms of what forms and created by what ways?To join the current interdisciplinary discussion on the issue of space, and to obtain new definition as well as insightful understanding of "space", this study explores the spatial phenomena in Internet. We hope that our findings would ultimately be also useful for contemporary architectural designers and scholars in their designs in the real world.As a preliminary exploration, the main objective of this study is to discover the elements involved in the creation/construction of Internet spaces and to examine the relationship between human participants and Internet spaces. In addition, this study also attempts to investigate whether participants from different academic disciplines define or experience Internet spaces in different ways, and to find what spatial elements of Internet they emphasize the most.In order to achieve a more comprehensive understanding of the spatial phenomena in Internet and to overcome the subjectivity of the members of the research team, the research design of this study was divided into two stages. At the first stage, we conducted literature review to study existing theories of space (which are based on observations and investigations of the physical world). At the second stage of this study, we recruited 8 Internet regular users to approach this topic from different point of views, and to see whether people with different academic training would define and experience Internet spaces differently.The results of this study reveal that the relationship between human participants and Internet spaces is different from that between human participants and physical spaces. In the physical world, physical elements of space must be established first; it then begins to be regarded as a place after interaction between/among human participants or interaction between human participants and the physical environment. In contrast, in Internet, a sense of place is first created through human interactions (or activities), Internet participants then begin to sense the existence of a space. Therefore, it seems that, among the many spatial elements of Internet we found, "interaction/reciprocity" Ñ either between/among human participants or between human participants and the computer interface Ð seems to be the most crucial element.In addition, another interesting result of this study is that verbal (linguistic) elements could provoke a sense of space in a degree higher than 2D visual representation and no less than 3D visual simulations. Nevertheless, verbal and 3D visual elements seem to work in different ways in terms of cognitive behaviors: Verbal elements provoke visual imagery and other sensory perceptions by "imagining" and then excite personal experiences of space; visual elements, on the other hand, provoke and excite visual experiences of space directly by "mapping".Finally, it was found that participants with different academic training did experience and define space differently. For example, when experiencing and analyzing Internet spaces, architecture designers, the creators of the physical world, emphasize the design of circulation and orientation, while participants with linguistics training focus more on subtle language usage. Visual designers tend to analyze the graphical elements of virtual spaces based on traditional painting theories; industrial designers, on the other hand, tend to treat these spaces as industrial products, emphasizing concept of user-center and the control of the computer interface.The findings of this study seem to add new information to our understanding of virtual space. It would be interesting for future studies to investigate how this information influences architectural designers in their real-world practices in this digital age. In addition, to obtain a fuller picture of Internet space, further research is needed to study the same issue by examining more Internet participants who have no formal linguistics and graphical training.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id eabb
authors Boeykens, St. Geebelen, B. and Neuckermans, H.
year 2002
title Design phase transitions in object-oriented modeling of architecture
source Connecting the Real and the Virtual - design e-ducation [20th eCAADe Conference Proceedings / ISBN 0-9541183-0-8] Warsaw (Poland) 18-20 September 2002, pp. 310-313
doi https://doi.org/10.52842/conf.ecaade.2002.310
summary The project IDEA+ aims to develop an “Integrated Design Environment for Architecture”. Its goal is providing a tool for the designer-architect that can be of assistance in the early-design phases. It should provide the possibility to perform tests (like heat or cost calculations) and simple simulations in the different (early) design phases, without the need for a fully detailed design or remodeling in a different application. The test for daylighting is already in development (Geebelen, to be published). The conceptual foundation for this design environment has been laid out in a scheme in which different design phases and scales are defined, together with appropriate tests at the different levels (Neuckermans, 1992). It is a translation of the “designerly” way of thinking of the architect (Cross, 1982). This conceptual model has been translated into a “Core Object Model” (Hendricx, 2000), which defines a structured object model to describe the necessary building model. These developments form the theoretical basis for the implementation of IDEA+ (both the data structure & prototype software), which is currently in progress. The research project addresses some issues, which are at the forefront of the architect’s interest while designing with CAAD. These are treated from the point of view of a practicing architect.
series eCAADe
email
last changed 2022/06/07 07:52

_id cf2011_p027
id cf2011_p027
authors Herssens, Jasmien; Heylighen Ann
year 2011
title A Framework of Haptic Design Parameters for Architects: Sensory Paradox Between Content and Representation
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 685-700.
summary Architects—like other designers—tend to think, know and work in a visual way. In design research, this way of knowing and working is highly valued as paramount to design expertise (Cross 1982, 2006). In case of architecture, however, it is not only a particular strength, but may as well be regarded as a serious weakness. The absence of non-visual features in traditional architectural spatial representations indicates how these are disregarded as important elements in conceiving space (Dischinger 2006). This bias towards vision, and the suppression of other senses—in the way architecture is conceived, taught and critiqued—results in a disappearance of sensory qualities (Pallasmaa 2005). Nevertheless, if architects design with more attention to non visual senses, they are able to contribute to more inclusive environments. Indeed if an environment offers a range of sensory triggers, people with different sensory capacities are able to navigate and enjoy it. Rather than implementing as many sensory triggers as possible, the intention is to make buildings and spaces accessible and enjoyable for more people, in line with the objective of inclusive design (Clarkson et al. 2007), also called Design for All or Universal Design (Ostroff 2001). Within this overall objective, the aim of our study is to develop haptic design parameters that support architects during design in paying more attention to the role of haptics, i.e. the sense of touch, in the built environment by informing them about the haptic implications of their design decisions. In the context of our study, haptic design parameters are defined as variables that can be decided upon by designers throughout the design process, and the value of which determines the haptic characteristics of the resulting design. These characteristics are based on the expertise of people who are congenitally blind, as they are more attentive to non visual information, and of professional caregivers working with them. The parameters do not intend to be prescriptive, nor to impose a particular method. Instead they seek to facilitate a more inclusive design attitude by informing designers and helping them to think differently. As the insights from the empirical studies with people born blind and caregivers have been reported elsewhere (Authors 2010), this paper starts by outlining the haptic design parameters resulting from them. Following the classification of haptics into active, dynamic and passive touch, the built environment unfolds into surfaces that can act as “movement”, “guiding” and/or “rest” plane. Furthermore design techniques are suggested to check the haptic qualities during the design process. Subsequently, the paper reports on a focus group interview/workshop with professional architects to assess the usability of the haptic design parameters for design practice. The architects were then asked to try out the parameters in the context of a concrete design project. The reactions suggest that the participating architects immediately picked up the underlying idea of the parameters, and recognized their relevance in relation to the design project at stake, but that their representation confronts us with a sensory paradox: although the parameters question the impact of the visual in architectural design, they are meant to be used by designers, who are used to think, know and work in a visual way.
keywords blindness, design parameters, haptics, inclusive design, vision
series CAAD Futures
email
last changed 2012/02/11 19:21

_id 5509
authors Koutamanis, Alexandros
year 1990
title Development of a computerized handbook of architectural plans
source Delft University of Technology
summary The dissertation investigates an approach to the development of visual / spatial computer representations for architectural purposes through the development of the computerized handbook of architectural plans (chap), a knowledge-based computer system capable of recognizing the metric properties of architectural plans. This investigation can be summarized as an introduction of computer vision to the computerization of architectural representations: chap represents an attempt to automate recognition of the most essential among conventional architectural drawings, floor plans. The system accepts as input digitized images of architectural plans and recognizes their spatial primitives (locations) and their spatial articulation on a variety of abstraction levels. The final output of chap is a description of the plan in terms of the grouping formations detected in its spatial articulation. The overall structure of the description is based on an analysis of its conformity to the formal rules of its “stylistic” context (which in the initial version of chap is classical architecture). Chapter 1 suggests that the poor performance of computerized architectural drawing and design systems is among others evidence of the necessity to computerize visual / spatial architectural representations. A recognition system such as chap offers comprehensive means for the investigation of a methodology for the development and use of such representations. Chapter 2 describes a fundamental task of chap: recognition of the position and shape of locations, the atomic parts of the description of an architectural plan in chap. This operation represents the final and most significant part of the first stage in processing an image input in machine environment. Chapter 3 moves to the next significant problem, recognition of the spatial arrangement of locations in an architectural plan, that is, recognition of grouping relationships that determine the subdivision of a plan into parts. In the absence of systematic and exhaustive typologic studies of classical architecture that would allow us to define a repertory of the location group types possible in classical architectural plans, Chapter 3 follows a bottom-up approach based on grouping relationships derived from elementary architectural knowledge and formalized with assistance from Gestalt theory and its antecedents. The grouping process described in Chapter 3 corresponds both in purpose and in structure to the derivation of a description of an image in computer vision [Marr 1982]. Chapter 4 investigates the well-formedness of the description of a classical architectural plan in an analytical manner: each relevant level (or sublevel) of the classical canon according to Tzonis & Lefaivre [1986] is transformed into a single group of criteria of well-formedness which is investigated independently. The hierarchical structure of the classical canon determines the coordination of these criteria into a sequence of cognitive filters which progressively analyses the correspondence of the descriptions derived as in Chapter 3 to the constraints of the canon. The methodology and techniques presented in the dissertation are primarily considered with respect to chap, a specific recognition system. The resulting specification of chap gives a measure of the use of such a system within the context of a computerized collection of architectural precedents and also presents several extensions to other areas of architecture. Although these extensions are not considered as verifiable claims, Chapter 5 describes some of their implications, including on the role of architectural drawing in computerized design systems, on architectural typologies, and on the nature and structure of generative systems in architecture.
series thesis:PhD
email
last changed 2003/02/12 22:37

_id 941c
id 941c
authors Maver, T.W. and Ellis, J.
year 1982
title Implementation of an Energy Model within a Multi-Disciplinary Practice
source Proceedings of CAD82, Brighton (Ed: A Pipes), Butterworth, 562-570
summary Implementation of computer software is concerned with trials of its robustness, relevance, and efficacy in the real-world, real-time context of design practice. This paper summarises the trials carried out within the Building Design Partnership of a dynamic energy model, ESP, developed by ABACUS at the University of Strathclyde. Over an 18 month period the program was used on 6 projects to address a variety of design problems over a variety of building types. The paper reports in outline on each of the six: four concerned with the need for a definite answer to a specific question, eg. ''will it overheat in summer"; two concerned to provide, on the one hand for the client, on the other hand for the design practice, paradigms for energy conscious design of hospitals and offices. The conclusions drawn have relevance to the takeup of CAD generally.
series other
email
last changed 2003/09/03 13:15

_id 4bae
authors Rasdorf, William J. and Kutay, Ali R.
year 1982
title Maintenance of Integrity During Concurrent Access in a Building Design Database
source Computer Aided Design. Butterworth Scientific Ltd., July, 1982. vol. 14: pp. 201-207. includes bibliography
summary This paper proposes a building design database model that insures database integrity in a highly flexible relational structure while supporting disciplinary and interdisciplinary concurrent use. The model strongly supports designer-database interaction by providing extremely versatile data access mechanisms and an associated concurrency control mechanism. Building design components are represented in terms of their location, their attribute values, and combinations of the two. Both the logical and physical database models are illustrated. The relational model is vital for achieving the greatest flexibility in representing and accessing building design data. Its standard relations are ideal for information representation. In addition, the operators provided by the model enable the engineer to readily restructure the database to support building design needs. This paper introduces a database structuring mechanism referred to as catalogs. Catalogs provide a highly versatile mechanism for accessing database information by grouping building components into data units called modules. The modules provide convenient access to multiple design entities. Also included is a protection relation that provides a concurrency control environment for the catalog relations. The module concept is particularly important in design because it enables the ad hoc groupings of data which are so often necessary to support the design process. The module is recommended as the level to which a locking concurrency control mechanism be applied. It is a small enough data unit to support concurrency for interdisciplinary design activities, yet not so small as to require extensive overhead in the concurrency control implementation. Two different modes of locking are recommended for the catalog relations of a building design database to achieve maximum concurrency and efficiency of access by designers
keywords database, concurrency, access, constraints management
series CADline
last changed 2003/06/02 13:58

_id 8a88
authors Anderson, David P.
year 1982
title Hidden Line Elimination in Projected Grid Surfaces
source ACM Transactions on Graphics. October, 1982. vol. 1: pp. 274-288 : ill. includes a short bibliography
summary The hidden line and hidden surface problems are simpler when restricted to special classes of objects. An example is the class of grid surfaces, that is, graphs of bivariate functions represented by their values on a set of grid points. Projected grid surfaces have geometric properties which permit hidden line or hidden surface elimination to be done more easily than in the general case. These properties are discussed in this paper, and an algorithm is given which exploits them
keywords algorithms, hidden lines, hidden surfaces, grids, computer graphics
series CADline
last changed 2003/06/02 10:24

_id 1b10
id 1b10
authors Bay, Joo-Hwa
year 2001
title Cognitive Biases - The case of tropical architecture
source Delft University of Technology
summary This dissertation investigates, i) How cognitive biases (or illusions) may lead to errors in design thinking, ii) Why architects use architectural precedents as heuristics despite such possible errors, and iii) Develops a design tool that can overcome this type of errors through the introduction of a rebuttal mechanism. The mechanism controls biases and improves accuracy in architectural thinking. // The research method applied is interdisciplinary. It employs knowledge from cognitive science, environmental engineering, and architectural theory. The case study approach is also used. The investigation is made in the case of tropical architecture. The investigation of architectural biases draws from work by A. Tversky and D. Kahneman in 1982 on “Heuristics and biases”. According to Tversky and Kahneman, the use of heuristics of representativeness (based on similarity) and availability (based on ease of recall and imaginability) for judgement of probability can result in cognitive biases of illusions of validity and biases due to imaginability respectively. This theory can be used analogically to understand how errors arise in the judgement of environmental behaviour anticipated from various spatial configurations, leading to designs with dysfunctional performances when built. Incomplete information, limited time, and human mental resources make design thinking in practice difficult and impossible to solve. It is not possible to analyse all possible alternative solutions, multiple contingencies, and multiple conflicting demands, as doing so will lead to combinatorial explosion. One of the ways to cope with the difficult design problem is to use precedents as heuristic devices, as shortcuts in design thinking, and at the risk of errors. This is done with analogical, pre-parametric, and qualitative means of thinking, without quantitative calculations. Heuristics can be efficient and reasonably effective, but may not always be good enough or even correct, because they can have associated cognitive biases that lead to errors. Several debiasing strategies are discussed, and one possibility is to introduce a rebuttal mechanism to refocus the designer’s thinking on the negative and opposite outcomes in his judgements, in order to debias these illusions. The research is carried out within the framework of design theory developed by the Design Knowledge System Research Centre, TUDelft. This strategy is tested with an experiment. The results show that the introduction of a rebuttal mechanism can debias and improve design judgements substantially in environmental control. The tool developed has possible applications in design practice and education, and in particular, in the designing of sustainable environments.
keywords Design bias; Design knowledge; Design rebuttal; Design Precedent; Pre-parametric design; Tropical architecture; Sustainability
series thesis:PhD
type normal paper
email
last changed 2006/05/28 07:42

_id 482a
authors Cole, Sam
year 1982
title A Microprocessor Revolution and the World Distribution of Income: A General Equilibrium Approach
source International Political Science Review. 1982. vol.3: pp. 434- 454 ; ill. includes bibliography
summary This article shows that even if the world economy is able to withstand and surmount the present world crisis, the combination of market forces and rapid technical change that would be the result of a microprocessor revolution will give rise to large shifts in the distribution of income within and between both rich and poor countries. Some developed and developing economies may be unable to join the move to new technologies. In a world governed by only economic forces, all countries, whether they choose to adopt new systems of production or not, will be affected. Indeed, whatever their degree of involvement, all countries are beginning to feel in varying degrees the chain reaction that reverberates through and between all sectors of their domestic and the world economies. To gain insights into interrelations between technological change and global markets, this article uses a special type of model -- a general equilibrium model -- that enables the study to focus on exactly these variables
keywords technology, economics
series CADline
last changed 1999/02/12 15:07

_id sigradi2006_e183a
id sigradi2006_e183a
authors Costa Couceiro, Mauro
year 2006
title La Arquitectura como Extensión Fenotípica Humana - Un Acercamiento Basado en Análisis Computacionales [Architecture as human phenotypic extension – An approach based on computational explorations]
source SIGraDi 2006 - [Proceedings of the 10th Iberoamerican Congress of Digital Graphics] Santiago de Chile - Chile 21-23 November 2006, pp. 56-60
summary The study describes some of the aspects tackled within a current Ph.D. research where architectural applications of constructive, structural and organization processes existing in biological systems are considered. The present information processing capacity of computers and the specific software development have allowed creating a bridge between two holistic nature disciplines: architecture and biology. The crossover between those disciplines entails a methodological paradigm change towards a new one based on the dynamical aspects of forms and compositions. Recent studies about artificial-natural intelligence (Hawkins, 2004) and developmental-evolutionary biology (Maturana, 2004) have added fundamental knowledge about the role of the analogy in the creative process and the relationship between forms and functions. The dimensions and restrictions of the Evo-Devo concepts are analyzed, developed and tested by software that combines parametric geometries, L-systems (Lindenmayer, 1990), shape-grammars (Stiny and Gips, 1971) and evolutionary algorithms (Holland, 1975) as a way of testing new architectural solutions within computable environments. It is pondered Lamarck´s (1744-1829) and Weismann (1834-1914) theoretical approaches to evolution where can be found significant opposing views. Lamarck´s theory assumes that an individual effort towards a specific evolutionary goal can cause change to descendents. On the other hand, Weismann defended that the germ cells are not affected by anything the body learns or any ability it acquires during its life, and cannot pass this information on to the next generation; this is called the Weismann barrier. Lamarck’s widely rejected theory has recently found a new place in artificial and natural intelligence researches as a valid explanation to some aspects of the human knowledge evolution phenomena, that is, the deliberate change of paradigms in the intentional research of solutions. As well as the analogy between genetics and architecture (Estévez and Shu, 2000) is useful in order to understand and program emergent complexity phenomena (Hopfield, 1982) for architectural solutions, also the consideration of architecture as a product of a human extended phenotype can help us to understand better its cultural dimension.
keywords evolutionary computation; genetic architectures; artificial/natural intelligence
series SIGRADI
email
last changed 2016/03/10 09:49

_id 4b04
authors De Wilde, W.P., Mollaert, M. and Buelinckx, H. (Ed.)
year 1983
title Proceedings of the International Conference eCAADe 1983
source Proceedings of the International Conference eCAADe [European Computer Aided Architectural Design Education] Brussels (Belgium) 1983
doi https://doi.org/10.52842/conf.ecaade.1983
summary In the beginning of the eighties, a few enthusiastic researchers, under the impulsion of Tom MAVER (director of the ABACUS group) and Rik SCHIJF (TH Delft) initiated a regular, if not formal, convention of people involved in the teaching of computer aided architectural design ECAADE in Europe. In 1982 a first meeting took place in Delft and, taking into consideration the member of attendants and the enthusiasm during this convention, it was decided that CAAD was definitely an important topic and that a more formal symposium was to be organised in the fall of 1983, in the University of Brussels.

The positive evolution of CAAD, not only in educational institutions, but also in professional practice is not surprising: it is to be considered in the global frame of technological and organisational revolution actually taking place. As will be read in the outstanding contribution of the participants it is not a mere choice of increased productivity which attracts the architects; the CAAD techniques also release then from a serious burden : the production of technical drawings and administrative paperwork!

series eCAADe
email
more http://wwwtw.vub.ac.be/ond/memc/Staff/Patrick.htm
last changed 2022/06/07 07:49

_id aed6
authors Hofstadter, Douglas R.
year 1982
title Metafont, Metamathematics, and Metaphysics
source 23, [38] p. : ill. Bloomington: Computer Science Department, Indiana University, december, 1982. Technical report No. 136. includes bibliography.--- Comments on Donald Knuth's Article 'The Concept of a Meta-Font.' --- See also Donald Knuth's 'The Concept of a Meta-Font,' Visible Language, (Winter 1982), vol. XVI, no. 1, pp. 3-27
summary It is argued that readers are likely to carry away from Donald Knuth's article 'The Concept of a Meta-Font' a falsely optimistic view of the extent to which the design of typefaces and letterforms can be mechanized through an approach depending on describing letterforms by specifying the settings of a large number of parameters. Through a comparison to mathematical logic, it is argued that no such set of parameters can capture the essence of any semantic category. Some different ways of thinking about the problem of the 'spirit' residing behind any letterform are suggested, connecting to current research issues in the field of artificial intelligence
keywords languages, AI, parametrization
series CADline
last changed 2003/06/02 13:58

_id 8c27
authors Kalay, Yehuda E.
year 1982
title Determining the Spatial Containment of a Point in General Polyhedra
source Computer graphics and Image Processing. 1982. vol. 19: pp. 303-334 : ill. includes bibliography. See also criticism and improvements in Orlowski, Marian
summary Determining the inclusion of a point in volume-enclosing polyhedra (shapes) in 3D space is, in principle, the extension of the well-known problem of determining the inclusion of a point in a polygon in 2D space. However, the extra degree of freedom makes 3D point-polyhedron containment analysis much more difficult to solve than the 2D point polygon problem, mainly because of the nonsequential ordering of the shape elements, which requires global shape data to be applied for resolving special cases. Two general O(n) algorithms for solving the problem by reducing the 3D case into the solvable 2D case are presented. The first algorithm, denoted 'the projection method,' is applicable to any planar- faced polyhedron, reducing the dimensionality by employing parallel projection to generate planar images of the shape faces, together with an image of the point being tested for inclusion. The containment relationship of these images is used to increment a global parity-counter when appropriate, representing an abstraction for counting the intersections between the surface of the shape and a halfline extending from the point to infinity. An 'inside' relationship is established when the parity-count is odd. Special cases (coincidence of the halfline with edges or vertices of the shape) are resolved by eliminating the coincidental elements and re-projecting the merged faces. The second algorithm, denoted 'the intersection method,' is applicable to any well- formed shape, including curved-surfaced ones. It reduces the dimensionality by intersecting the polygonal trace of the shape surface at the plane of intersection, which is tested for containing the trace of the point in the plane, directly establishing the overall 3D containment relationship. A particular O(n) implementation of the 2D point-in-polygon inclusion algorithm, which is used for solving the problem once reduced in dimensionality, is also presented. The presentation is complemented by discussions of the problems associated with point-polyhedron relationship determination in general, and comparative analysis of the two particular algorithms presented
keywords geometric modeling, point inclusion, polygons, polyhedra, computational geometry, algorithms, search, B-rep
series CADline
email
last changed 2003/06/02 10:24

_id c6e3
authors Kirsch, Uri
year 1982
title Approximation Concepts for Optimum Structural Design
source 25 p. Pittsburgh, PA: Design Research Center, CMU, December, 1982. DRC-12-12-82. includes bibliography
summary When elastic behavior is assumed, the optimum structural design problem is stated in an implicit nonlinear programming form. Approximate behavior models are introduced to obtain explicit constraint functions, essential in many practical design problems. For the displacement analysis formulation, some possible explicit approximations of the nodal displacements in terms of the design variables are reviewed. It is shown that any approximation of the displacements will lead to a solution which does not necessarily satisfy equilibrium. Methods for solving the approximate problem are discussed. Using the force analysis formulation, it is shown that explicit approximations of the redundant forces will lead to solutions which do not necessarily satisfy compatibility. Some simplified models of the explicit problem are presented and solution methods are discussed
keywords structures, design, engineering, optimization, civil engineering, approximation
series CADline
last changed 1999/02/12 15:08

_id 2c13
authors Knuth, Donald E.
year 1982
title The Concept of a Meta-Font
source Visible Language. Winter, 1982. vol. XVI: pp. 3-27 : ill. includes bibliography
summary A single drawing of a single letter reveals only a small part of what was in the designer's mind when that letter was drawn. But when precise instructions are given about how to make such a drawing, the intelligence of that letter can be captured in a way that permits us to obtain an infinite variety of related letters from the same specification. Instead of merely describing a single letter, such instructions explain how that letter would change its shape if other parameters of the design were changed. Thus an entire font of letters and other symbols can be specified so that each character adapts itself to varying conditions in an appropriate way. Initial experiments with a precise language for pen motions suggest strongly that the font designer of the future should not simply design isolated alphabets; the challenge will be to explain exactly how each design should adapt itself gracefully to a wide range of changes in the specification. This paper gives examples of a meta-font and explains the changeable parameters in its design
keywords programming, computer graphics, parametrization
series CADline
last changed 2003/06/02 13:58

_id 0095
authors Kowalski, T.J. and Thomas, Donald E.
year 1982
title The VLSI Design Automation Assistant : First Step
source 5 p. Design Research Center, CMU, December, 1982. DRC-18-57-82. includes bibliography
summary This paper describes an approach to VLSI design synthesis using both knowledge-based expert systems and data and control flow analysis. The authors are concerned with design synthesis as it proceeds from an algorithmic description of a VLSI system to a list of technology-independent registers, operators, data paths, and control signals. This paper discusses the development of the Design Automation Assistant from its first interviews with expert VLSI designers to its current prototype state. Four designs of a microcomputer are presented along with the changes in the knowledge base that created those designs
keywords design, automation, integrated circuits, synthesis, expert systems, systems, knowledge base
series CADline
last changed 2003/06/02 13:58

_id c361
authors Logan, Brian S.
year 1986
title Representing the Structure of Design Problems
source Computer-Aided Architectural Design Futures [CAAD Futures Conference Proceedings / ISBN 0-408-05300-3] Delft (The Netherlands), 18-19 September 1985, pp. 158-170
summary In recent years several experimental CAD systems have emerged which, focus specifically on the structure of design problems rather than on solution generation or appraisal (Sussman and Steele, 1980; McCallum, 1982). However, the development of these systems has been hampered by the lack of an adequate theoretical basis. There is little or no argument as to what the statements comprising these models actually mean, or on the types of operations that should be provided. This chapter describes an attempt to develop a semantically adequate basis for a model of the structure of design problems and presents a representation of this model in formal logic.
series CAAD Futures
last changed 1999/04/03 17:58

_id fafa
authors Marr, David
year 1982
title Vision: A Computational Investigation into the Human Representation and Processing of Visual Information
source New York, NY: W.H. Freeman and Company
summary Marr's demonstrations that retinal receptive field geometry could be derived by Fourier transformation of spatial frequency sensitivity data, that edges and contours could be detected by finding zero crossings in the light gradient by taking the Laplacian or second directional derivative, that excitatory and inhibitory receptive fields could be constructed from "DOG" functions (the difference of two Gaussians), and that the visual system used a two-dimensional convolution integral with a Gaussian prefilter as an operator for bandwidth optimation on the retinal light distribution, were more powerful than anything that had been seen up to that time. It was as if vision research suddenly acquired its own Principia Mathematica, or perhaps General Relativity Theory, in terms of the new explanatory power Marr's theories provided. Truly an extraordinary book from an extraordinary thinker in the area of perception, vision, and the brain.
series other
last changed 2003/04/23 15:14

For more results click below:

this is page 0show page 1show page 2show page 3HOMELOGIN (you are user _anon_18234 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002