CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 232

_id a3f5
authors Zandi-Nia, Abolfazl
year 1992
title Topgene: An artificial Intelligence Approach to a Design Process
source Delft University of Technology
summary This work deals with two architectural design (AD) problems at the topological level and in presence of the social norms community, privacy, circulation-cost, and intervening opportunity. The first problem concerns generating a design with respect to a set of above mentioned norms, and the second problem requires evaluation of existing designs with respect to the same set of norms. Both problems are based on the structural-behavioral relationship in buildings. This work has challenged above problems in the following senses: (1) A working system, called TOPGENE (The TOpological Pattern GENErator) has been developed. (2) Both problems may be vague and may lack enough information in their statement. For example, an AD in the presence of the social norms requires the degrees of interactions between the location pairs in the building. This information is not always implicitly available, and must be explicated from the design data. (3) An AD problem at topological level is intractable with no fast and efficient algorithm for its solution. To reduce the search efforts in the process of design generation, TOPGENE uses a heuristic hill climbing strategy that takes advantage of domain specific rules of thumbs to choose a path in the search space of a design. (4) TOPGENE uses the Q-analysis method for explication of hidden information, also hierarchical clustering of location-pairs with respect to their flow generation potential as a prerequisite information for the heuristic reasoning process. (5) To deal with a design of a building at topological level TOPGENE takes advantage of existing graph algorithms such as path-finding and planarity testing during its reasoning process. This work also presents a new efficient algorithm for keeping track of distances in a growing graph. (6) This work also presents a neural net implementation of a special case of the design generation problem. This approach is based on the Hopfield model of neural networks. The result of this approach has been used test TOPGENE approach in generating designs. A comparison of these two approaches shows that the neural network provides mathematically more optimal designs, while TOPGENE produces more realistic designs. These two systems may be integrated to create a hybrid system.
series thesis:PhD
last changed 2003/02/12 22:37

_id 6e99
authors Hoffer, Erin Rae
year 1992
title Creating the Electronic Design Studio: Development of a Heterogeneous Networked Environment at Harvard's Graduate School of Design
doi https://doi.org/10.52842/conf.ecaade.1992.225
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 225-240
summary The migration of design education to reliance on computer-based techniques requires new ways of thinking about environments which can effectively support a diverse set of activities. Both from a spatial standpoint and a computing resource standpoint, design studios must be inevitably reconfigured to support new tools and reflect new ways of communicating. At Harvard's GSD, a commitment to incorporating computer literacy as a fundamental component of design education enables us to confront these issues through the implementation of a heterogeneous network imbedded in an electronic design environment. This evolving prototype of a new design studio, its development and its potential, will be the subject of this paper. A new style design environment is built upon an understanding of traditional techniques, and layered with an awareness of new tools and methods. Initially we borrow from existing metaphors which govern our interpretation of the way designers work. Next we seek to extend our thinking to include allied or related metaphors such as the library metaphor which informs collections of software and data, or the laboratory metaphor which informs workspace groupings, or the transportation metaphor which informs computer-based communications such as electronic mail or bulletin boards, or the utility services metaphor which informs the provision of network services and equipment. Our evaluation of this environment is based on direct feedback from its users, both faculty and students, and on subjective observation of the qualitative changes in communication which occur between and among these groups and individuals. Ultimately, the network must be judged as a framework for learning and evaluation, and its success depends both on its ability to absorb our existing metaphors for the process of design, and to prefigure the emerging metaphors to be envisioned in the future.

series eCAADe
last changed 2022/06/07 07:50

_id 0719
authors Shiffer, M.J.
year 1992
title Towards a collaborative planning system
source Environment and Planning B, Volume 19, 1992, pp. 709-722
summary This article begins by exploring the problem of combining the elements of group cognition, access to media, and access to tools into a holistic planning process. It then discusses a way in which technology can be used to help combine these activities by incorporating graphical interfaces, associative information structuring, and computer-supported collaborative work into a microcomputer-based Collaborative Planning System (CPS). Methods for the development of a CPS are proposed and two systems are explored as examples. It is concluded that increased access to relevant information, aided by the implementation of a CPS, can ultimately lead to greater communication amongst participants in a group planning situation. This will ultimately have a positive effect on the quality of plans and decisions.
series journal paper
last changed 2003/04/23 15:50

_id b4c4
authors Carrara, G., Fioravanti, A. and Novembri, G.
year 2000
title A framework for an Architectural Collaborative Design
doi https://doi.org/10.52842/conf.ecaade.2000.057
source Promise and Reality: State of the Art versus State of Practice in Computing for the Design and Planning Process [18th eCAADe Conference Proceedings / ISBN 0-9523687-6-5] Weimar (Germany) 22-24 June 2000, pp. 57-60
summary The building industry involves a larger number of disciplines, operators and professionals than other industrial processes. Its peculiarity is that the products (building objects) have a number of parts (building elements) that does not differ much from the number of classes into which building objects can be conceptually subdivided. Another important characteristic is that the building industry produces unique products (de Vries and van Zutphen, 1992). This is not an isolated situation but indeed one that is spreading also in other industrial fields. For example, production niches have proved successful in the automotive and computer industries (Carrara, Fioravanti, & Novembri, 1989). Building design is a complex multi-disciplinary process, which demands a high degree of co-ordination and co-operation among separate teams, each having its own specific knowledge and its own set of specific design tools. Establishing an environment for design tool integration is a prerequisite for network-based distributed work. It was attempted to solve the problem of efficient, user-friendly, and fast information exchange among operators by treating it simply as an exchange of data. But the failure of IGES, CGM, PHIGS confirms that data have different meanings and importance in different contexts. The STandard for Exchange of Product data, ISO 10303 Part 106 BCCM, relating to AEC field (Wix, 1997), seems to be too complex to be applied to professional studios. Moreover its structure is too deep and the conceptual classifications based on it do not allow multi-inheritance (Ekholm, 1996). From now on we shall adopt the BCCM semantic that defines the actor as "a functional participant in building construction"; and we shall define designer as "every member of the class formed by designers" (architects, engineers, town-planners, construction managers, etc.).
keywords Architectural Design Process, Collaborative Design, Knowledge Engineering, Dynamic Object Oriented Programming
series eCAADe
email
more http://www.uni-weimar.de/ecaade/
last changed 2022/06/07 07:55

_id ddss9209
id ddss9209
authors De Gelder, J.T. and Lucardie, G.L.
year 1993
title Knowledge and data modelling in cad/cam applications
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Modelling knowledge and data in CAD/CAM applications is complex because different goals and contexts have to be taken into account. This complexity makes particular demands upon representation formalisms. Today many modelling tools are based on record structures. By analyzing the requirements for a product model of a portal structure in steel, this paper shows that in many situations record structures are not well suited as a representation formalism for storing knowledge and data in CAD/CAM applications. This is illustrated by performing a knowledge-level analysis of the knowledge and data generated in the design and manufacturing process of a portal structure in steel.
series DDSS
last changed 2003/08/07 16:36

_id 6ea4
authors Eastman, C.M.
year 1992
title A Data Model Analysis of Modularity and Extensibility in Building Databases
source Building and Environment, Vol 27, No: 2, pp. 135-148
summary This paper uses data modeling techniques to define how database schemas for an intelligent integrated architectural CAD system can be made extensible. It reviews the product data modeling language EDM, then applies it to define a part of an architectural data model. Extensions are then investigated, regarding how users could integrate various design-specific packages into a uniquely configured system. Both, extension by substituting one technology for another and by adding a new evaluation application, are considered. Data modeling allows specification of a CAD database and identification of the kind of modularization that will work and what problems may arise.''
series journal paper
email
last changed 2003/04/23 15:14

_id 7ce5
authors Gal, Shahaf
year 1992
title Computers and Design Activities: Their Mediating Role in Engineering Education
source Sociomedia, ed. Edward Barret. MIT Press
summary Sociomedia: With all the new words used to describe electronic communication (multimedia, hypertext, cyberspace, etc.), do we need another one? Edward Barrett thinks we do; hence, he coins the term "sociomedia." It is meant to displace a computing economy in which technicity is hypostasized over sociality. Sociomedia, a compilation of twenty-five articles on the theory, design and practice of educational multimedia and hypermedia, attempts to re-value the communicational face of computing. Value, of course, is "ultimately a social construct." As such, it has everything to do with knowledge, power, education and technology. The projects discussed in this book represent the leading edge of electronic knowledge production in academia (not to mention major funding) and are determining the future of educational media. For these reasons, Sociomedia warrants close inspection. Barrett's introduction sets the tone. For him, designing computer media involves hardwiring a mechanism for the social construction of knowledge (1). He links computing to a process of social and communicative interactivity for constructing and desseminating knowledge. Through a mechanistic mapping of the university as hypercontext (a huge network that includes classrooms as well as services and offices), Barrett models intellectual work in such a way as to avoid "limiting definitions of human nature or human development." Education, then, can remain "where it should be--in the human domain (public and private) of sharing ideas and information through the medium of language." By leaving education in a virtual realm (where we can continue to disagree about its meaning and execution), it remains viral, mutating and contaminating in an intellectually healthy way. He concludes that his mechanistic model, by means of its reductionist approach, preserves value (7). This "value" is the social construction of knowledge. While I support the social orientation of Barrett's argument, discussions of value are related to power. I am not referring to the traditional teacher-student power structure that is supposedly dismantled through cooperative and constructivist learning strategies. The power to be reckoned with in the educational arena is foundational, that which (pre)determines value and the circulation of knowledge. "Since each of you reading this paragraph has a different perspective on the meaning of 'education' or 'learning,' and on the processes involved in 'getting an education,' think of the hybris in trying to capture education in a programmable function, in a displayable object, in a 'teaching machine'" (7). Actually, we must think about that hybris because it is, precisely, what informs teaching machines. Moreover, the basic epistemological premises that give rise to such productions are too often assumed. In the case of instructional design, the episteme of cognitive sciences are often taken for granted. It is ironic that many of the "postmodernists" who support electronic hypertextuality seem to have missed Jacques Derrida's and Michel Foucault's "deconstructions" of the epistemology underpinning cognitive sciences (if not of epistemology itself). Perhaps it is the glitz of the technology that blinds some users (qua developers) to the belief systems operating beneath the surface. Barrett is not guilty of reactionary thinking or politics; he is, in fact, quite in line with much American deconstructive and postmodern thinking. The problem arises in that he leaves open the definitions of "education," "learning" and "getting an education." One cannot engage in the production of new knowledge without orienting its design, production and dissemination, and without negotiating with others' orientations, especially where largescale funding is involved. Notions of human nature and development are structural, even infrastructural, whatever the medium of the teaching machine. Although he addresses some dynamics of power, money and politics when he talks about the recession and its effects on the conference, they are readily visible dynamics of power (3-4). Where does the critical factor of value determination, of power, of who gets what and why, get mapped onto a mechanistic model of learning institutions? Perhaps a mapping of contributors' institutions, of the funding sources for the projects showcased and for participation in the conference, and of the disciplines receiving funding for these sorts of projects would help visualize the configurations of power operative in the rising field of educational multimedia. Questions of power and money notwithstanding, Barrett's introduction sets the social and textual thematics for the collection of essays. His stress on interactivity, on communal knowledge production, on the society of texts, and on media producers and users is carried foward through the other essays, two of which I will discuss. Section I of the book, "Perspectives...," highlights the foundations, uses and possible consequences of multimedia and hypertextuality. The second essay in this section, "Is There a Class in This Text?," plays on the robust exchange surrounding Stanley Fish's book, Is There a Text in This Class?, which presents an attack on authority in reading. The author, John Slatin, has introduced electronic hypertextuality and interaction into his courses. His article maps the transformations in "the content and nature of work, and the workplace itself"-- which, in this case, is not industry but an English poetry class (25). Slatin discovered an increase of productive and cooperative learning in his electronically- mediated classroom. For him, creating knowledge in the electronic classroom involves interaction between students, instructors and course materials through the medium of interactive written discourse. These interactions lead to a new and persistent understanding of the course materials and of the participants' relation to the materials and to one another. The work of the course is to build relationships that, in my view, constitute not only the meaning of individual poems, but poetry itself. The class carries out its work in the continual and usually interactive production of text (31). While I applaud his strategies which dismantle traditional hierarchical structures in academia, the evidence does not convince me that the students know enough to ask important questions or to form a self-directing, learning community. Stanley Fish has not relinquished professing, though he, too, espouses the indeterminancy of the sign. By the fourth week of his course, Slatin's input is, by his own reckoning, reduced to 4% (39). In the transcript of the "controversial" Week 6 exchange on Gertrude Stein--the most disliked poet they were discussing at the time (40)--we see the blind leading the blind. One student parodies Stein for three lines and sums up his input with "I like it." Another, finds Stein's poetry "almost completey [sic] lacking in emotion or any artistic merit" (emphasis added). On what grounds has this student become an arbiter of "artistic merit"? Another student, after admitting being "lost" during the Wallace Steven discussion, talks of having more "respect for Stevens' work than Stein's" and adds that Stein's poetry lacks "conceptual significance[, s]omething which people of varied opinion can intelligently discuss without feeling like total dimwits...." This student has progressed from admitted incomprehension of Stevens' work to imposing her (groundless) respect for his work over Stein's. Then, she exposes her real dislike for Stein's poetry: that she (the student) missed the "conceptual significance" and hence cannot, being a person "of varied opinion," intelligently discuss it "without feeling like [a] total dimwit." Slatin's comment is frightening: "...by this point in the semester students have come to feel increasingly free to challenge the instructor" (41). The students that I have cited are neither thinking critically nor are their preconceptions challenged by student-governed interaction. Thanks to the class format, one student feels self-righteous in her ignorance, and empowered to censure. I believe strongly in student empowerment in the classroom, but only once students have accrued enough knowledge to make informed judgments. Admittedly, Slatin's essay presents only partial data (there are six hundred pages of course transcripts!); still, I wonder how much valuable knowledge and metaknowledge was gained by the students. I also question the extent to which authority and professorial dictature were addressed in this course format. The power structures that make it possible for a college to require such a course, and the choice of texts and pedagogy, were not "on the table." The traditional professorial position may have been displaced, but what took its place?--the authority of consensus with its unidentifiable strong arm, and the faceless reign of software design? Despite Slatin's claim that the students learned about the learning process, there is no evidence (in the article) that the students considered where their attitudes came from, how consensus operates in the construction of knowledge, how power is established and what relationship they have to bureaucratic insitutions. How do we, as teaching professionals, negotiate a balance between an enlightened despotism in education and student-created knowledge? Slatin, and other authors in this book, bring this fundamental question to the fore. There is no definitive answer because the factors involved are ultimately social, and hence, always shifting and reconfiguring. Slatin ends his article with the caveat that computerization can bring about greater estrangement between students, faculty and administration through greater regimentation and control. Of course, it can also "distribute authority and power more widely" (50). Power or authority without a specific face, however, is not necessarily good or just. Shahaf Gal's "Computers and Design Activities: Their Mediating Role in Engineering Education" is found in the second half of the volume, and does not allow for a theory/praxis dichotomy. Gal recounts a brief history of engineering education up to the introduction of Growltiger (GT), a computer-assisted learning aid for design. He demonstrates GT's potential to impact the learning of engineering design by tracking its use by four students in a bridge-building contest. What his text demonstrates clearly is that computers are "inscribing and imaging devices" that add another viewpoint to an on-going dialogue between student, teacher, earlier coursework, and other teaching/learning tools. The less proficient students made a serious error by relying too heavily on the technology, or treating it as a "blueprint provider." They "interacted with GT in a way that trusted the data to represent reality. They did not see their interaction with GT as a negotiation between two knowledge systems" (495). Students who were more thoroughly informed in engineering discourses knew to use the technology as one voice among others--they knew enough not simply to accept the input of the computer as authoritative. The less-advanced students learned a valuable lesson from the competition itself: the fact that their designs were not able to hold up under pressure (literally) brought the fact of their insufficient knowledge crashing down on them (and their bridges). They also had, post factum, several other designs to study, especially the winning one. Although competition and comparison are not good pedagogical strategies for everyone (in this case the competitors had volunteered), at some point what we think we know has to be challenged within the society of discourses to which it belongs. Students need critique in order to learn to push their learning into auto-critique. This is what is lacking in Slatin's discussion and in the writings of other avatars of constructivist, collaborative and computer-mediated pedagogies. Obviously there are differences between instrumental types of knowledge acquisition and discoursive knowledge accumulation. Indeed, I do not promote the teaching of reading, thinking and writing as "skills" per se (then again, Gal's teaching of design is quite discursive, if not dialogic). Nevertheless, the "soft" sciences might benefit from "bridge-building" competitions or the re-institution of some forms of agonia. Not everything agonistic is inhuman agony--the joy of confronting or creating a sound argument supported by defensible evidence, for example. Students need to know that soundbites are not sound arguments despite predictions that electronic writing will be aphoristic rather than periodic. Just because writing and learning can be conceived of hypertextually does not mean that rigor goes the way of the dinosaur. Rigor and hypertextuality are not mutually incompatible. Nor is rigorous thinking and hard intellectual work unpleasurable, although American anti-intellectualism, especially in the mass media, would make it so. At a time when the spurious dogmatics of a Rush Limbaugh and Holocaust revisionist historians circulate "aphoristically" in cyberspace, and at a time when knowledge is becoming increasingly textualized, the role of critical thinking in education will ultimately determine the value(s) of socially constructed knowledge. This volume affords the reader an opportunity to reconsider knowledge, power, and new communications technologies with respect to social dynamics and power relationships.
series other
last changed 2003/04/23 15:14

_id 130d
authors Hoinkes, R. and Mitchell, R.
year 1994
title Playing with Time - Continuous Temporal Mapping Strategies for Interactive Environments
source 6th Canadian GIS Conference, (Ottawa Natura Resources Canada), pp. 318-329
summary The growing acceptance of GIS technology has had far- reaching effects on many fields of research. The recent developments in the area of dynamic and temporal GIS open new possibilities within the realm of historical research where temporal relationship analysis is as important as spatial relationship analysis. While topological structures have had wide use in spatial GIS and have been the subject of most temporal GIS endeavours, the different demands of many of these temporally- oriented analytic processes questions the choice of the topological direction. In the fall of 1992 the Montreal Research Group (MRG) of the Canadian Centre for Architecture mounted an exhibition dealing with the development of the built environment in 18th- century Montreal. To aid in presenting the interpretive messages of their data, the MRG worked with the Centre for Landscape Research (CLR) to incorporate the interactive capabilities of the CLR's PolyTRIM research software with the MRG's data base to produce a research tool as well as a public- access interactive display. The interactive capabilities stemming from a real- time object- oriented structure provided an excellent environment for both researchers and the public to investigate the nature of temporal changes in such aspects as landuse, ethnicity, and fortifications of the 18th century city. This paper describes the need for interactive real- time GIS in such temporal analysis projects and the underlying need for object- oriented vs. topologically structured data access strategies to support them.
series other
last changed 2003/04/23 15:14

_id ab4d
authors Huang, Tao-Kuang, Degelman, Larry O., and Larsen, Terry R.
year 1992
title A Visualization Model for Computerized Energy Evaluation During the Conceptual Design Stage (ENERGRAPH)
doi https://doi.org/10.52842/conf.acadia.1992.195
source Mission - Method - Madness [ACADIA Conference Proceedings / ISBN 1-880250-01-2] 1992, pp. 195-206
summary Energy performance is a crucial step toward responsible design. Currently there are many tools that can be applied to reach this goal with reasonable accuracy. Often times, however, major flaws are not discovered until the final stage of design when it is too late to change. Not only are existing simulation models complicated to apply at the conceptual design stage, but energy principles and their applications are also abstract and hard to visualize. Because of the lack of suitable tools to visualize energy analysis output, energy conservation concepts fail to be integrated into the building design. For these reasons, designers tend not to apply energy conservation concepts at the early design stage. However, since computer graphics is a new phase of visual communication in design process, the above problems might be solved properly through a computerized graphical interface in the conceptual design stage.

The research described in this paper is the result of exploring the concept of using computer graphics to support energy efficient building designs. It focuses on the visualization of building energy through a highly interactive graphical interface in the early design stage.

series ACADIA
email
last changed 2022/06/07 07:50

_id 56e9
authors Huang, Tao-Kuang
year 1992
title A Graphical Feedback Model for Computerized Energy Analysis during the Conceptual Design Stage
source Texas A&M University
summary During the last two decades, considerable effort has been placed on the development of building design analysis tools. Architects and designers have begun to take advantage of computers to generate and examine design alternatives. However, because it has been difficult to adapt computer technologies to the visual orientation of the building designer, the majority of computer applications have been limited to numerical analysis and office automation tasks. Only recently, because of advances in hardware and software techniques, computers have entered into a new phase in the development of architectural design. haveters are now able to interactively display graphics solutions to architectural related problems, which is fundamental to the design process. The majority of research programs in energy efficient design have sharpened people's understanding of energy principles and their application of those principles. Energy conservation concepts, however, have not been widely used. A major problem in the implementation of these principles is that energy principles their applications are abstract, hard to visualize and separated from the architectural design process. Furthermore, one aspect of energy analysis may contain thousands of pieces of numerical information which often leads to confusion on the part of designers. If these difficulties can be overcome, it would bring a great benefit to the advancement of energy conservation concepts. This research explores the concept of an integrated computer graphics program to support energy efficient design. It focuses on (1) the integration of energy efficiently and architectural design, and (2) the visualization of building energy use through graphical interfaces during the conceptual design stage. It involves (1) the discussion of frameworks of computer-aided architectural design and computer-aided energy efficient building design, and (2) the development of an integrated computer prototype program with a graphical interface that helps the designer create building layouts, analyze building energy interactively and receive visual feedbacks dynamically. The goal is to apply computer graphics as an aid to visualize the effects of energy related decisions and therefore permit the designer to visualize and understand energy conservation concepts in the conceptual phase of architectural design.
series thesis:PhD
last changed 2003/02/12 22:37

_id caadria2024_365
id caadria2024_365
authors Lahtinen, Aaro, Gardner, Nicole, Ramos Jaime, Cristina and Yu, Kuai
year 2024
title Visualising Sydney's Urban Green: A Web Interface for Monitoring Vegetation Coverage between 1992 and 2022 using Google Earth Engine
doi https://doi.org/10.52842/conf.caadria.2024.2.515
source Nicole Gardner, Christiane M. Herr, Likai Wang, Hirano Toshiki, Sumbul Ahmad Khan (eds.), ACCELERATED DESIGN - Proceedings of the 29th CAADRIA Conference, Singapore, 20-26 April 2024, Volume 2, pp. 515–524
summary With continued population growth and urban expansion, the severity of environmental concerns within cities is likely to increase without proper urban ecosystem monitoring and management. Despite this, limited efforts have been made to effectively communicate the ecological value of urban vegetation to Architecture, Engineering and Construction (AEC) professionals concerned with mitigating these effects and improving urban liveability. In response, this research project proposes a novel framework for identifying and conveying historical changes to vegetation coverage within the Greater Sydney area between 1992 and 2022. The cloud-based geo-spatial analysis platform, Google Earth Engine (GEE), was used to construct an accurate land cover classification of Landsat imagery, allowing the magnitude, spatial configuration, and period of vegetation loss to be promptly identified. The outcomes of this analysis are represented through an intuitive web platform that facilitates a thorough understanding of the complex relationships between anthropogenic activities and vegetation coverage. A key finding indicated that recent developments in the Blacktown area had directly contributed to heightened land surface temperature, suggesting a reformed approach to urban planning is required to address climatic concerns appropriately. The developed web interface provides a unique method for AEC professionals to assess the effectiveness of past planning strategies, encouraging a multi-disciplinary approach to urban ecosystem management.
keywords Urban Vegetation, Web Interface, Landsat Imagery, Land Cover Classification, Google Earth Engine
series CAADRIA
email
last changed 2024/11/17 22:05

_id 181b
authors Liou, Shuenn-Ren
year 1992
title A computer-based framework for analyzing and deriving the morphological structure of architectural designs
source University of Michigan
summary An approach to the acquisition and utilization of knowledge about the morphological structure of notable orthogonal building plans and other two-dimensional compositions is formulated and tested. This approach consists of two levels of abstraction within which the analysis and comparison of existing designs and the derivation of new designs can be undertaken systematically and efficiently. Specifically, the morphological structure of orthogonal building plans and other two-dimensional compositions is conceived as a language defined by shape grammar and architectural grammar corresponding to the geometric and spatial structures of the compositions. Lines constitute the shape grammar and walls and columns the architectural grammar. A computer program named ANADER is designed and implemented using the C++ object-oriented language to describe feasible compositions. It is argued that the gap between morphological analysis and synthesis is bridged partially because the proposed framework facilitates systematic comparisons of the morphological structures of two-dimensional orthogonal compositions and provides insight into the form-making process used to derive them. As an analytical system, the framework contributes to the generation of new and the assessment of existing morphological knowledge. Specifically, it is demonstrated that it is feasible to specify an existing architectural design by a set of universal rule schemata and the sequence of their application. As a generative system, the framework allows many of the tasks involved in the derivation of two-dimensional orthogonal compositions to be carried out. As well, it promotes the use of analytical results. In conclusion, it is argued that the proposed computer-based framework will provide the research and the educator with increasing opportunities for addressing persistent architectural questions in new ways. Of particular interest to this author are questions concerning the decision-making activities involved in form- and space-making as well as the description, classification, and derivation of architecutural form and space. It is suggested that, at least in reference to the cases examined, but probably also in reference to many other morphological classes, these and other related questions can be addressed systematically, efficiently, and fruitfully by using the proposed framework.  
series thesis:PhD
last changed 2003/02/12 22:37

_id ddss9208
id ddss9208
authors Lucardie, G.L.
year 1993
title A functional approach to realizing decision support systems in technical regulation management for design and construction
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Technical building standards defining the quality of buildings, building products, building materials and building processes aim to provide acceptable levels of safety, health, usefulness and energy consumption. However, the logical consistency between these goals and the set of regulations produced to achieve them is often hard to identify. Not only the large quantities of highly complex and frequently changing building regulations to be met, but also the variety of user demands and the steadily increasing technical information on (new) materials, products and buildings have produced a very complex set of knowledge and data that should be taken into account when handling technical building regulations. Integrating knowledge technology and database technology is an important step towards managing the complexity of technical regulations. Generally, two strategies can be followed to integrate knowledge and database technology. The main emphasis of the first strategy is on transferring data structures and processing techniques from one field of research to another. The second approach is concerned exclusively with the semantic structure of what is contained in the data-based or knowledge-based system. The aim of this paper is to show that the second or knowledge-level approach, in particular the theory of functional classifications, is more fundamental and more fruitful. It permits a goal-directed rationalized strategy towards analysis, use and application of regulations. Therefore, it enables the reconstruction of (deep) models of regulations, objects and of users accounting for the flexibility and dynamics that are responsible for the complexity of technical regulations. Finally, at the systems level, the theory supports an effective development of a new class of rational Decision Support Systems (DSS), which should reduce the complexity of technical regulations and restore the logical consistency between the goals of technical regulations and the technical regulations themselves.
series DDSS
last changed 2003/08/07 16:36

_id aa6d
authors Nichols, Foster Jr., Canete, Isabel J. and Tuladhar, Sagun
year 1992
title Designing for Pedestrians : A CAD-Network Analysis Approach
source New York: John Wiley & Sons, 1992. pp. 379-398 : ill. includes a short bibliography
summary Microcomputer techniques have been developed that combine CAD drawings with transportation network analysis software that uses spreadsheets and stand-alone programs activated from the DOS operating system. The CAD feature simplifies and improves the methods used to design pedestrian circulation facilities and evaluate the impact of new development on existing pedestrian flows. Through the use of customized software, the need for manual data entry is reduced, and the graphical display of analysis results in most intermediate steps in the process are automated. Three hypothetical case studies are presented, concentrating on proposed pedestrian circulation improvements at Penn Station, New York
keywords evaluation, networks, management, CAD, analysis, applications, planning, transportation, prediction, simulation, CAD
series CADline
last changed 2003/06/02 13:58

_id 2c22
authors O'Neill, Michael J.
year 1992
title Neural Network Simulation as a Computer- Aided design Tool For Predicting Wayfinding Performance
source New York: John Wiley & Sons, 1992. pp. 347-366 : ill. includes bibliography
summary Complex public facilities such as libraries, hospitals, and governmental buildings often present problems to users who must find their way through them. Research shows that difficulty in wayfinding has costs in terms of time, money, public safety, and stress that results from being lost. While a wide range of architectural research supports the notion that ease of wayfinding should be a criterion for good design, architects have no method for evaluating how well their building designs will support the wayfinding task. People store and retrieve information about the layout of the built environment in a knowledge representation known as the cognitive map. People depend on the information stored in the cognitive map to find their way through buildings. Although there are numerous simulations of the cognitive map, the mechanisms of these models are not constrained by what is known about the neurophysiology of the brain. Rather, these models incorporate search mechanisms that act on semantically encoded information about the environment. In this paper the author describes the evaluation and application of an artificial neural network simulation of the cognitive map as a means of predicting wayfinding behavior in buildings. This simulation is called NAPS-PC (Network Activity Processing Simulator--PC version). This physiologically plausible model represents knowledge about the layout of the environment through a network of inter-connected processing elements. The performance of NAPS-PC was evaluated against actual human wayfinding performance. The study found that the simulation generated behavior that matched the performance of human participants. After the validation, NAPS-PC was modified so that it could read environmental information directly from AutoCAD (a popular micro-computer-based CAD software package) drawing files, and perform 'wayfinding' tasks based on that environmental information. This prototype tool, called AutoNet, is conceptualized as a means of allowing designers to predict the wayfinding performance of users in a building before it is actually built
keywords simulation, cognition, neural networks, evaluation, floor plans, applications, wayfinding, layout, building
series CADline
last changed 2003/06/02 13:58

_id eaff
authors Shaviv, Edna and Kalay, Yehuda E.
year 1992
title Combined Procedural and Heuristic Method to Energy Conscious Building Design and Evaluation
source New York: John Wiley & Sons, 1992. pp. 305-325 : ill. includes bibliography
summary This paper describes a methodology that combines both procedural and heuristic methods by means of integrating a simulation model with a knowledge based system (KBS) for supporting all phases of energy conscious design and evaluation. The methodology is based on partitioning the design process into discrete phases and identifying the informational characteristics of each phase, as far as energy conscious design is concerned. These informational characteristics are expressed in the form of design variables (parameters) and the relationships between them. The expected energy performance of a design alternative is evaluated by a combination of heuristic and procedural methods, and the context-sensitive application of default values, when necessary. By virtue of combining knowledge based evaluations with procedural ones, this methodology allows for testing the applicability of heuristic rules in non-standard cases,Ô h)0*0*0*°° ÔŒ thereby improving the predictable powers of the evaluation
keywords design process, evaluation, energy, analysis, synthesis, integration, architecture, knowledge base, heuristics, simulation
series CADline
email
last changed 2003/06/02 10:24

_id avocaad_2001_20
id avocaad_2001_20
authors Shen-Kai Tang
year 2001
title Toward a procedure of computer simulation in the restoration of historical architecture
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In the field of architectural design, “visualization¨ generally refers to some media, communicating and representing the idea of designers, such as ordinary drafts, maps, perspectives, photos and physical models, etc. (Rahman, 1992; Susan, 2000). The main reason why we adopt visualization is that it enables us to understand clearly and to control complicated procedures (Gombrich, 1990). Secondly, the way we get design knowledge is more from the published visualized images and less from personal experiences (Evans, 1989). Thus the importance of the representation of visualization is manifested.Due to the developments of computer technology in recent years, various computer aided design system are invented and used in a great amount, such as image processing, computer graphic, computer modeling/rendering, animation, multimedia, virtual reality and collaboration, etc. (Lawson, 1995; Liu, 1996). The conventional media are greatly replaced by computer media, and the visualization is further brought into the computerized stage. The procedure of visual impact analysis and assessment (VIAA), addressed by Rahman (1992), is renewed and amended for the intervention of computer (Liu, 2000). Based on the procedures above, a great amount of applied researches are proceeded. Therefore it is evident that the computer visualization is helpful to the discussion and evaluation during the design process (Hall, 1988, 1990, 1992, 1995, 1996, 1997, 1998; Liu, 1997; Sasada, 1986, 1988, 1990, 1993, 1997, 1998). In addition to the process of architectural design, the computer visualization is also applied to the subject of construction, which is repeatedly amended and corrected by the images of computer simulation (Liu, 2000). Potier (2000) probes into the contextual research and restoration of historical architecture by the technology of computer simulation before the practical restoration is constructed. In this way he established a communicative mode among archeologists, architects via computer media.In the research of restoration and preservation of historical architecture in Taiwan, many scholars have been devoted into the studies of historical contextual criticism (Shi, 1988, 1990, 1991, 1992, 1995; Fu, 1995, 1997; Chiu, 2000). Clues that accompany the historical contextual criticism (such as oral information, writings, photographs, pictures, etc.) help to explore the construction and the procedure of restoration (Hung, 1995), and serve as an aid to the studies of the usage and durability of the materials in the restoration of historical architecture (Dasser, 1990; Wang, 1998). Many clues are lost, because historical architecture is often age-old (Hung, 1995). Under the circumstance, restoration of historical architecture can only be proceeded by restricted pictures, written data and oral information (Shi, 1989). Therefore, computer simulation is employed by scholars to simulate the condition of historical architecture with restricted information after restoration (Potier, 2000). Yet this is only the early stage of computer-aid restoration. The focus of the paper aims at exploring that whether visual simulation of computer can help to investigate the practice of restoration and the estimation and evaluation after restoration.By exploring the restoration of historical architecture (taking the Gigi Train Station destroyed by the earthquake in last September as the operating example), this study aims to establish a complete work on computer visualization, including the concept of restoration, the practice of restoration, and the estimation and evaluation of restoration.This research is to simulate the process of restoration by computer simulation based on visualized media (restricted pictures, restricted written data and restricted oral information) and the specialized experience of historical architects (Potier, 2000). During the process of practicing, communicates with craftsmen repeatedly with some simulated alternatives, and makes the result as the foundation of evaluating and adjusting the simulating process and outcome. In this way we address a suitable and complete process of computer visualization for historical architecture.The significance of this paper is that we are able to control every detail more exactly, and then prevent possible problems during the process of restoration of historical architecture.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id fd02
authors Tsou, Jin-Yeu
year 1992
title Using conceptual modelling and an object-oriented environment to support building cost control during early design
source College of Architecture and Urban Planning, University of Michigan
summary This research investigated formal information modelling techniques and the object-oriented knowledge representation on the domain of building cost control during early design stages. The findings contribute to an understanding of the advantages and disadvantages of applying formal modelling techniques to the analysis of architectural problems and the representation of domain knowledge in an object-oriented environment. In this study, information modelling techniques were reviewed, formal information analysis was performed, a conceptual model based on the cost control problem domain was created, a computational model based on the object-oriented approach was developed, a mechanism to support information broadcasting for representing interrelationships was implemented, and an object-oriented cost analysis system for early design (OBCIS) was demonstrated. The conceptual model, based on the elemental proposition analysis of NIAM, supports a formal approach for analyzing the problem domain; the analysis results are represented by high-level graphical notations, based on the AEC Building System Model, to visually display the information framework of the domain. The conceptual model provides an intermediate step between the system designer's view of the domain and the internal representation of the implementation platform. The object-oriented representation provides extensive data modelling abilities to help system designers intuitively represent the semantics of the problem domain. The object-oriented representation also supports more structured and integrated modules than conventional programming approaches. Although there are many advantages to applying this technique to represent the semantics of cost control knowledge, there are several issues which need to be considered: no single satisfactory classification method can be directly applied; object-oriented systems are difficult to learn; and designing reusable classes is difficult. The dependency graph and information broadcasting implemented in this research is an attempt to represent the interrelationships between domain objects. The mechanism allows users to explicitly define the interrelationships, based on semantic requirements, among domain objects. In the conventional approach, these relationships are directly interpreted by system designers and intertwined into the programming code. There are several issues which need to be studied further: indirect dependency relationship, conflict resolution, and request-update looping based on least-commitment approach.
series thesis:PhD
email
last changed 2003/02/12 22:37

_id cbed
authors Yakubu, G.S.
year 1994
title Maximising the Benefits of CAD Systems in Architectural Education
doi https://doi.org/10.52842/conf.ecaade.1994.x.u8n
source The Virtual Studio [Proceedings of the 12th European Conference on Education in Computer Aided Architectural Design / ISBN 0-9523687-0-6] Glasgow (Scotland) 7-10 September 1994, p. 228
summary The positive impact of Computer Aided Design (CAD) in professional architectural practice has been in focus in recent times but relatively little has been written on its significance in the education of the contemporary architect. It is common knowledge that the profession of architecture is currently undergoing enormous strains as it battles to keep abreast of trends and developments in a period of series of rapid advancement in science, technology and management (RIBA, 1992). Whilst attempts are being made to redress the shortcomings of the profession in the above context, the requirements for architectural education are yet to forge a coherent strategy for the implementation of CAD/IT in the curriculum of schools of architecture. In almost every other field, including engineering, medicine and the humanities, computing application to problem-solving and decision-making is seen as a way forward as we move into 21st century. Architectural education must integrate CAD/IT into the teaching of core modules that give the architect distinctive competence: studio design. That is one of the best ways of doing justice to the education of the architect of today and the future. Some approaches to the teaching of CAD in schools of architecture have been touched upon in the recent past. Building upon this background as well as an understanding of the nature of design teaching/learning, this paper examines ways of maximising the benefits of CAD systems in architectural education and of bringing computer aided designing into the studio not only to enhance design thinking and creativity but also to support interactive processes. In order to maximise or optimise any function, one approach is to use the hard systems methodology which utilises analytic, analogic and iconic models to show the effect of those factors which are significant for the purposes being considered. The other approach is to use the soft systems methodology in which the analysis encompasses the concept of a human activity system as a means of improving a situation. The use of soft systems methodology is considered more appropriate for dealing with the problem of design which is characterised by a flux of interacting events and ideas that unroll through time. The paper concludes that the main impediment to maximising the benefits of CAD systems in architectural education is not only the inappropriate definition of the objectives for the implementation of CAD education but also that the control subsystems are usually ill-structured and relatively poorly defined. Schools must attempt to define a coherent and consistent policy on the use of CAD systems as an integral part of studio design and evolve an in-house strategic and operational controls that enable the set objectives to be met. Furthermore, it is necessary to support the high level of productivity from CAD systems with a more efficient management system, especially in dealing with communication, data sharing via relational database, co-ordination and integration. Finally, the use of soft systems methodology is recommended as the way forward to optimising CAD systems in design education as it would provide continuous improvements while maintaining their productive value.

series eCAADe
last changed 2022/06/07 07:50

_id avocaad_2001_09
id avocaad_2001_09
authors Yu-Tung Liu, Yung-Ching Yeh, Sheng-Cheng Shih
year 2001
title Digital Architecture in CAD studio and Internet-based competition
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary Architectural design has been changing because of the vast and creative use of computer in different ways. From the viewpoint of designing itself, computer has been used as drawing tools in the latter phase of design (Mitchell 1977; Coyne et al. 1990), presentation and simulation tools in the middle phase (Liu and Bai 2000), and even critical media which triggers creative thinking in the very early phase (Maher et al. 2000; Liu 1999; Won 1999). All the various roles that computer can play have been adopted in a number of professional design corporations and so-called computer-aided design (CAD) studio in schools worldwide (Kvan 1997, 2000; Cheng 1998). The processes and outcomes of design have been continuously developing to capture the movement of the computer age. However, from the viewpoint of social-cultural theories of architecture, the evolvement of design cannot be achieved solely by designers or design processes. Any new idea of design can be accepted socially, culturally and historically only under one condition: The design outcomes could be reviewed and appreciated by critics in the field at the time of its production (Csikszentmihalyi 1986, 1988; Schon and Wiggins 1992; Liu 2000). In other words, aspects of design production (by designers in different design processes) are as critical as those of design appreciation (by critics in different review processes) in the observation of the future trends of architecture.Nevertheless, in the field of architectural design with computer and Internet, that is, so-called computer-aided design computer-mediated design, or internet-based design, most existing studies pay more attentions to producing design in design processes as mentioned above. Relatively few studies focus on how critics act and how they interact with designers in the review processes. Therefore, this study intends to investigate some evolving phenomena of the interaction between design production and appreciation in the environment of computer and Internet.This paper takes a CAD studio and an Internet-based competition as examples. The CAD studio includes 7 master's students and 2 critics, all from the same countries. The Internet-based competition, held in year 2000, includes 206 designers from 43 counties and 26 critics from 11 countries. 3 students and the 2 critics in the CAD studio are the competition participating designers and critics respectively. The methodological steps are as follows: 1. A qualitative analysis: observation and interview of the 3 participants and 2 reviewers who join both the CAD studio and the competition. The 4 analytical criteria are the kinds of presenting media, the kinds of supportive media (such as verbal and gesture/facial data), stages of the review processes, and interaction between the designer and critics. The behavioral data are acquired by recording the design presentation and dialogue within 3 months. 2. A quantitative analysis: statistical analysis of the detailed reviewing data in the CAD studio and the competition. The four 4 analytical factors are the reviewing time, the number of reviewing of the same project, the comparison between different projects, and grades/comments. 3. Both the qualitative and quantitative data are cross analyzed and discussed, based on the theories of design thinking, design production/appreciation, and the appreciative system (Goodman 1978, 1984).The result of this study indicates that the interaction between design production and appreciation during the review processes could differ significantly. The review processes could be either linear or cyclic due to the influences from the kinds of media, the environmental discrepancies between studio and Internet, as well as cognitive thinking/memory capacity. The design production and appreciation seem to be more linear in CAD studio whereas more cyclic in the Internet environment. This distinction coincides with the complementary observations of designing as a linear process (Jones 1970; Simon 1981) or a cyclic movement (Schon and Wiggins 1992). Some phenomena during the two processes are also illustrated in detail in this paper.This study is merely a starting point of the research in design production and appreciation in the computer and network age. The future direction of investigation is to establish a theoretical model for the interaction between design production and appreciation based on current findings. The model is expected to conduct using revised protocol analysis and interviews. The other future research is to explore how design computing creativity emerge from the process of producing and appreciating.
series AVOCAAD
email
last changed 2005/09/09 10:48

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 11HOMELOGIN (you are user _anon_896944 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002