CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 247

_id 5b34
authors Langendorf, R.
year 1992
title The 1990's: information systems and computer visualization for urban design, planning and management
source Environment and Planning B: Planning and Design 19, pp. 723-738
summary Contributed by Susan Pietsch (spietsch@arch.adelaide.edu.au)
keywords 3D City Modeling, Development Control, Design Control
series other
last changed 2001/06/04 20:38

_id 338b
authors Bishop, I. D.
year 1992
title Data integration for visualization: application to decision support
source AURSIA '92: Proceedings of the Twentieth Annual International Conference of the Australian Urban and Regional Information Systems Association Incorporated, Gold Coast, Australia, 25-27 November 1992, Australian Urban and Regional Information Systems Association Incorporated, ACT, Australia, pp. 74-80
summary Contributed by Susan Pietsch (spietsch@arch.adelaide.edu.au)
keywords 3D City Modeling, Development Control, Design Control
series other
last changed 2001/06/04 20:27

_id cef3
authors Bridges, Alan H.
year 1992
title Computing and Problem Based Learning at Delft University of Technology Faculty of Architecture
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 289-294
doi https://doi.org/10.52842/conf.ecaade.1992.289
summary Delft University of Technology, founded in 1842, is the oldest and largest technical university in the Netherlands. It provides education for more than 13,000 students in fifteen main subject areas. The Faculty of Architecture, Housing, Urban Design and Planning is one of the largest faculties of the DUT with some 2000 students and over 500 staff members. The course of study takes four academic years: a first year (Propaedeuse) and a further three years (Doctoraal) leading to the "ingenieur" qualification. The basic course material is delivered in the first two years and is taken by all students. The third and fourth years consist of a smaller number of compulsory subjects in each of the department's specialist areas together with a wide range of option choices. The five main subject areas the students may choose from for their specialisation are Architecture, Building and Project Management, Building Technology, Urban Design and Planning, and Housing.

The curriculum of the Faculty has been radically revised over the last two years and is now based on the concept of "Problem-Based Learning". The subject matter taught is divided thematically into specific issues that are taught in six week blocks. The vehicles for these blocks are specially selected and adapted case studies prepared by teams of staff members. These provide a focus for integrating specialist subjects around a studio based design theme. In the case of second year this studio is largely computer-based: many drawings are produced by computer and several specially written computer applications are used in association with the specialist inputs.

This paper describes the "block structure" used in second year, giving examples of the special computer programs used, but also raises a number of broader educational issues. Introduction of the block system arose as a method of curriculum integration in response to difficulties emerging from the independent functioning of strong discipline areas in the traditional work groups. The need for a greater level of selfdirected learning was recognised as opposed to the "passive information model" of student learning in which the students are seen as empty vessels to be filled with knowledge - which they are then usually unable to apply in design related contexts in the studio. Furthermore, the value of electives had been questioned: whilst enabling some diversity of choice, they may also be seen as diverting attention and resources from the real problems of teaching architecture.

series eCAADe
email
last changed 2022/06/07 07:54

_id caadria2010_042
id caadria2010_042
authors Celento, David
year 2010
title Open-source, parametric architecture to propagate hyper-dense, sustainable urban communities: parametric urban dwellings for the experience economy
source Proceedings of the 15th International Conference on Computer Aided Architectural Design Research in Asia / Hong Kong 7-10 April 2010, pp. 443-452
doi https://doi.org/10.52842/conf.caadria.2010.443
summary Rapid developments in societal, technological, and natural systems suggest profound changes ahead if research in panarchical systems (Holling, 2001) is to be believed. Panarchy suggests that systems, both natural and man-made, rise to the point of vulnerability then fail due to disruptive forces in a process of ‘creative destruction.’ This sequence allows for radical, and often unpredictable, renewal. Pressing sustainability concerns, burgeoning urban growth, and emergent ‘green manufacturing’ laws, suggest that future urban dwellings are headed toward Gladwell’s ‘tipping point’ (2002). Hyper-dense, sustainable, urban communities that employ open-source standards, parametric software, and web-based configurators are the new frontier for venerable visions. Open-source standards will permit the design, manufacture, and sale of highly diverse, inter-operable components to create compact urban living environments that are technologically sophisticated, sustainable, and mobile. These mass-customised dwellings, akin to branded consumer goods, will address previous shortcomings for prefabricated, mobile dwellings by stimulating consumer desire in ways that extend the arguments of both Joseph Pine (1992) and Anna Klingman (2007). Arguments presented by authors Makimoto and Manners (1997) – which assert that the adoption of digital and mobile technologies will create large-scale societal shifts – will be extended with several solutions proposed.
keywords Mass customisation; urban dwellings; open source standards; parametric design; sustainability
series CAADRIA
email
last changed 2022/06/07 07:55

_id 4857
authors Escola Tecnica Superior D'arquitectura de Barcelona (Ed.)
year 1992
title CAAD Instruction: The New Teaching of an Architect?
source eCAADe Conference Proceedings / Barcelona (Spain) 12-14 November 1992, 551 p.
doi https://doi.org/10.52842/conf.ecaade.1992
summary The involvement of computer graphic systems in the transmission of knowledge in the areas of urban planning and architectural design will bring a significant change to the didactic programs and methods of those schools which have decided to adopt these new instruments. Workshops of urban planning and architectural design will have to modify their structures, and teaching teams will have to revise their current programs. Some european schools and faculties of architecture have taken steps in this direction. Others are willing to join them.

This process is only delayed by the scarcity of material resources, and by the slowness with which a sufficient number of teachers are adopting these methods.

ECAADE has set out to analyze the state of this issue during its next conference, and it will be discussed from various points of view. From this confrontation of ideas will come, surely, the guidelines for progress in the years to come.

The different sessions will be grouped together following these four themes:

(A.) Multimedia and Course Work / State of the art of the synthesis of graphical and textual information favored by new available multimedia computer programs. Their repercussions on academic programs. (B.) The New Design Studio / Physical characteristics, data concentration and accessibility of a computerized studio can be better approached in a computerized workshop. (C.) How to manage the new education system / Problems and possibilities raised, from the practical and organizational points of view, of architectural education by the introduction of computers in the classrooms. (D.) CAAI. Formal versus informal structure / How will the traditional teaching structure be affected by the incidence of these new systems in which the access to knowledge and information can be obtained in a random way and guided by personal and subjective criteria.

series eCAADe
email
last changed 2022/06/07 07:49

_id 68c8
authors Flemming, U., Coyne, R. and Fenves, S. (et al.)
year 1994
title SEED: A Software Environment to Support the Early Phases in Building Design
source Proceeding of IKM '94, Weimar, Germany, pp. 5-10
summary The SEED project intends to develop a software environment that supports the early phases in building design (Flemming et al., 1993). The goal is to provide support, in principle, for the preliminary design of buildings in all aspects that can gain from computer support. This includes using the computer not only for analysis and evaluation, but also more actively for the generation of designs, or more accurately, for the rapid generation of design representations. A major motivation for the development of SEED is to bring the results of two multi-generational research efforts focusing on `generative' design systems closer to practice: 1. LOOS/ABLOOS, a generative system for the synthesis of layouts of rectangles (Flemming et al., 1988; Flemming, 1989; Coyne and Flemming, 1990; Coyne, 1991); 2. GENESIS, a rule-based system that supports the generation of assemblies of 3-dimensional solids (Heisserman, 1991; Heisserman and Woodbury, 1993). The rapid generation of design representations can take advantage of special opportunities when it deals with a recurring building type, that is, a building type dealt with frequently by the users of the system. Design firms - from housing manufacturers to government agencies - accumulate considerable experience with recurring building types. But current CAD systems capture this experience and support its reuse only marginally. SEED intends to provide systematic support for the storing and retrieval of past solutions and their adaptation to similar problem situations. This motivation aligns aspects of SEED closely with current work in Artificial Intelligence that focuses on case-based design (see, for example, Kolodner, 1991; Domeshek and Kolodner, 1992; Hua et al., 1992).
series other
email
last changed 2003/04/23 15:14

_id 7ce5
authors Gal, Shahaf
year 1992
title Computers and Design Activities: Their Mediating Role in Engineering Education
source Sociomedia, ed. Edward Barret. MIT Press
summary Sociomedia: With all the new words used to describe electronic communication (multimedia, hypertext, cyberspace, etc.), do we need another one? Edward Barrett thinks we do; hence, he coins the term "sociomedia." It is meant to displace a computing economy in which technicity is hypostasized over sociality. Sociomedia, a compilation of twenty-five articles on the theory, design and practice of educational multimedia and hypermedia, attempts to re-value the communicational face of computing. Value, of course, is "ultimately a social construct." As such, it has everything to do with knowledge, power, education and technology. The projects discussed in this book represent the leading edge of electronic knowledge production in academia (not to mention major funding) and are determining the future of educational media. For these reasons, Sociomedia warrants close inspection. Barrett's introduction sets the tone. For him, designing computer media involves hardwiring a mechanism for the social construction of knowledge (1). He links computing to a process of social and communicative interactivity for constructing and desseminating knowledge. Through a mechanistic mapping of the university as hypercontext (a huge network that includes classrooms as well as services and offices), Barrett models intellectual work in such a way as to avoid "limiting definitions of human nature or human development." Education, then, can remain "where it should be--in the human domain (public and private) of sharing ideas and information through the medium of language." By leaving education in a virtual realm (where we can continue to disagree about its meaning and execution), it remains viral, mutating and contaminating in an intellectually healthy way. He concludes that his mechanistic model, by means of its reductionist approach, preserves value (7). This "value" is the social construction of knowledge. While I support the social orientation of Barrett's argument, discussions of value are related to power. I am not referring to the traditional teacher-student power structure that is supposedly dismantled through cooperative and constructivist learning strategies. The power to be reckoned with in the educational arena is foundational, that which (pre)determines value and the circulation of knowledge. "Since each of you reading this paragraph has a different perspective on the meaning of 'education' or 'learning,' and on the processes involved in 'getting an education,' think of the hybris in trying to capture education in a programmable function, in a displayable object, in a 'teaching machine'" (7). Actually, we must think about that hybris because it is, precisely, what informs teaching machines. Moreover, the basic epistemological premises that give rise to such productions are too often assumed. In the case of instructional design, the episteme of cognitive sciences are often taken for granted. It is ironic that many of the "postmodernists" who support electronic hypertextuality seem to have missed Jacques Derrida's and Michel Foucault's "deconstructions" of the epistemology underpinning cognitive sciences (if not of epistemology itself). Perhaps it is the glitz of the technology that blinds some users (qua developers) to the belief systems operating beneath the surface. Barrett is not guilty of reactionary thinking or politics; he is, in fact, quite in line with much American deconstructive and postmodern thinking. The problem arises in that he leaves open the definitions of "education," "learning" and "getting an education." One cannot engage in the production of new knowledge without orienting its design, production and dissemination, and without negotiating with others' orientations, especially where largescale funding is involved. Notions of human nature and development are structural, even infrastructural, whatever the medium of the teaching machine. Although he addresses some dynamics of power, money and politics when he talks about the recession and its effects on the conference, they are readily visible dynamics of power (3-4). Where does the critical factor of value determination, of power, of who gets what and why, get mapped onto a mechanistic model of learning institutions? Perhaps a mapping of contributors' institutions, of the funding sources for the projects showcased and for participation in the conference, and of the disciplines receiving funding for these sorts of projects would help visualize the configurations of power operative in the rising field of educational multimedia. Questions of power and money notwithstanding, Barrett's introduction sets the social and textual thematics for the collection of essays. His stress on interactivity, on communal knowledge production, on the society of texts, and on media producers and users is carried foward through the other essays, two of which I will discuss. Section I of the book, "Perspectives...," highlights the foundations, uses and possible consequences of multimedia and hypertextuality. The second essay in this section, "Is There a Class in This Text?," plays on the robust exchange surrounding Stanley Fish's book, Is There a Text in This Class?, which presents an attack on authority in reading. The author, John Slatin, has introduced electronic hypertextuality and interaction into his courses. His article maps the transformations in "the content and nature of work, and the workplace itself"-- which, in this case, is not industry but an English poetry class (25). Slatin discovered an increase of productive and cooperative learning in his electronically- mediated classroom. For him, creating knowledge in the electronic classroom involves interaction between students, instructors and course materials through the medium of interactive written discourse. These interactions lead to a new and persistent understanding of the course materials and of the participants' relation to the materials and to one another. The work of the course is to build relationships that, in my view, constitute not only the meaning of individual poems, but poetry itself. The class carries out its work in the continual and usually interactive production of text (31). While I applaud his strategies which dismantle traditional hierarchical structures in academia, the evidence does not convince me that the students know enough to ask important questions or to form a self-directing, learning community. Stanley Fish has not relinquished professing, though he, too, espouses the indeterminancy of the sign. By the fourth week of his course, Slatin's input is, by his own reckoning, reduced to 4% (39). In the transcript of the "controversial" Week 6 exchange on Gertrude Stein--the most disliked poet they were discussing at the time (40)--we see the blind leading the blind. One student parodies Stein for three lines and sums up his input with "I like it." Another, finds Stein's poetry "almost completey [sic] lacking in emotion or any artistic merit" (emphasis added). On what grounds has this student become an arbiter of "artistic merit"? Another student, after admitting being "lost" during the Wallace Steven discussion, talks of having more "respect for Stevens' work than Stein's" and adds that Stein's poetry lacks "conceptual significance[, s]omething which people of varied opinion can intelligently discuss without feeling like total dimwits...." This student has progressed from admitted incomprehension of Stevens' work to imposing her (groundless) respect for his work over Stein's. Then, she exposes her real dislike for Stein's poetry: that she (the student) missed the "conceptual significance" and hence cannot, being a person "of varied opinion," intelligently discuss it "without feeling like [a] total dimwit." Slatin's comment is frightening: "...by this point in the semester students have come to feel increasingly free to challenge the instructor" (41). The students that I have cited are neither thinking critically nor are their preconceptions challenged by student-governed interaction. Thanks to the class format, one student feels self-righteous in her ignorance, and empowered to censure. I believe strongly in student empowerment in the classroom, but only once students have accrued enough knowledge to make informed judgments. Admittedly, Slatin's essay presents only partial data (there are six hundred pages of course transcripts!); still, I wonder how much valuable knowledge and metaknowledge was gained by the students. I also question the extent to which authority and professorial dictature were addressed in this course format. The power structures that make it possible for a college to require such a course, and the choice of texts and pedagogy, were not "on the table." The traditional professorial position may have been displaced, but what took its place?--the authority of consensus with its unidentifiable strong arm, and the faceless reign of software design? Despite Slatin's claim that the students learned about the learning process, there is no evidence (in the article) that the students considered where their attitudes came from, how consensus operates in the construction of knowledge, how power is established and what relationship they have to bureaucratic insitutions. How do we, as teaching professionals, negotiate a balance between an enlightened despotism in education and student-created knowledge? Slatin, and other authors in this book, bring this fundamental question to the fore. There is no definitive answer because the factors involved are ultimately social, and hence, always shifting and reconfiguring. Slatin ends his article with the caveat that computerization can bring about greater estrangement between students, faculty and administration through greater regimentation and control. Of course, it can also "distribute authority and power more widely" (50). Power or authority without a specific face, however, is not necessarily good or just. Shahaf Gal's "Computers and Design Activities: Their Mediating Role in Engineering Education" is found in the second half of the volume, and does not allow for a theory/praxis dichotomy. Gal recounts a brief history of engineering education up to the introduction of Growltiger (GT), a computer-assisted learning aid for design. He demonstrates GT's potential to impact the learning of engineering design by tracking its use by four students in a bridge-building contest. What his text demonstrates clearly is that computers are "inscribing and imaging devices" that add another viewpoint to an on-going dialogue between student, teacher, earlier coursework, and other teaching/learning tools. The less proficient students made a serious error by relying too heavily on the technology, or treating it as a "blueprint provider." They "interacted with GT in a way that trusted the data to represent reality. They did not see their interaction with GT as a negotiation between two knowledge systems" (495). Students who were more thoroughly informed in engineering discourses knew to use the technology as one voice among others--they knew enough not simply to accept the input of the computer as authoritative. The less-advanced students learned a valuable lesson from the competition itself: the fact that their designs were not able to hold up under pressure (literally) brought the fact of their insufficient knowledge crashing down on them (and their bridges). They also had, post factum, several other designs to study, especially the winning one. Although competition and comparison are not good pedagogical strategies for everyone (in this case the competitors had volunteered), at some point what we think we know has to be challenged within the society of discourses to which it belongs. Students need critique in order to learn to push their learning into auto-critique. This is what is lacking in Slatin's discussion and in the writings of other avatars of constructivist, collaborative and computer-mediated pedagogies. Obviously there are differences between instrumental types of knowledge acquisition and discoursive knowledge accumulation. Indeed, I do not promote the teaching of reading, thinking and writing as "skills" per se (then again, Gal's teaching of design is quite discursive, if not dialogic). Nevertheless, the "soft" sciences might benefit from "bridge-building" competitions or the re-institution of some forms of agonia. Not everything agonistic is inhuman agony--the joy of confronting or creating a sound argument supported by defensible evidence, for example. Students need to know that soundbites are not sound arguments despite predictions that electronic writing will be aphoristic rather than periodic. Just because writing and learning can be conceived of hypertextually does not mean that rigor goes the way of the dinosaur. Rigor and hypertextuality are not mutually incompatible. Nor is rigorous thinking and hard intellectual work unpleasurable, although American anti-intellectualism, especially in the mass media, would make it so. At a time when the spurious dogmatics of a Rush Limbaugh and Holocaust revisionist historians circulate "aphoristically" in cyberspace, and at a time when knowledge is becoming increasingly textualized, the role of critical thinking in education will ultimately determine the value(s) of socially constructed knowledge. This volume affords the reader an opportunity to reconsider knowledge, power, and new communications technologies with respect to social dynamics and power relationships.
series other
last changed 2003/04/23 15:14

_id 56e9
authors Huang, Tao-Kuang
year 1992
title A Graphical Feedback Model for Computerized Energy Analysis during the Conceptual Design Stage
source Texas A&M University
summary During the last two decades, considerable effort has been placed on the development of building design analysis tools. Architects and designers have begun to take advantage of computers to generate and examine design alternatives. However, because it has been difficult to adapt computer technologies to the visual orientation of the building designer, the majority of computer applications have been limited to numerical analysis and office automation tasks. Only recently, because of advances in hardware and software techniques, computers have entered into a new phase in the development of architectural design. haveters are now able to interactively display graphics solutions to architectural related problems, which is fundamental to the design process. The majority of research programs in energy efficient design have sharpened people's understanding of energy principles and their application of those principles. Energy conservation concepts, however, have not been widely used. A major problem in the implementation of these principles is that energy principles their applications are abstract, hard to visualize and separated from the architectural design process. Furthermore, one aspect of energy analysis may contain thousands of pieces of numerical information which often leads to confusion on the part of designers. If these difficulties can be overcome, it would bring a great benefit to the advancement of energy conservation concepts. This research explores the concept of an integrated computer graphics program to support energy efficient design. It focuses on (1) the integration of energy efficiently and architectural design, and (2) the visualization of building energy use through graphical interfaces during the conceptual design stage. It involves (1) the discussion of frameworks of computer-aided architectural design and computer-aided energy efficient building design, and (2) the development of an integrated computer prototype program with a graphical interface that helps the designer create building layouts, analyze building energy interactively and receive visual feedbacks dynamically. The goal is to apply computer graphics as an aid to visualize the effects of energy related decisions and therefore permit the designer to visualize and understand energy conservation concepts in the conceptual phase of architectural design.
series thesis:PhD
last changed 2003/02/12 22:37

_id 11b6
authors Kalmychkov, Vitaly A. and Smolyaninov, Alexander V.
year 1992
title Design of Object-Oriented Data Visualization System
source East-West International Conference on Human-Computer Interaction: Proceedings of the EWHCI'92 1992 pp. 463-470
summary The report is devoted to the data visualization system design and implementation, which provides the means for design of the image of the user's numeric information on the personal computer. The problems of design, architecture and operation of data visualization system which provides to user convenient means for constructing the numeric information image of required type is considered. Image constructing is executed by means of required sizes fields placing and filling of them by necessary content (coordinates system, graphs, inscriptions). User's interface with instrument system is object-oriented: after object (field or its content) choice user can manipulate of it, executing only those operations, that are determined for it as object of appointed function. Ergonomical and comfortable constructing is ensured by careful coordinated system of possible actions on each of image constructing stage and supported by icons menu and textual menu.
series other
last changed 2002/07/07 16:01

_id 5a7c
authors Schneiderman, B.
year 1992
title Designing the User Interface. Strategies for Effective Human-Computer Interaction
source Reading, Mass. etc, Addison-Wesley
summary Ben Shneiderman again provides a complete, current, and authoritative introduction to user-interface design. Students will learn practical techniques and guidelines needed to develop good systems designs - systems with interfaces the typical user can understand, predict, and control. This third edition features new chapters on the World Wide Web, information visualization, and computer-supported cooperative work. It contains expanded and earlier coverage of development methodologies, evaluation techniques, and user-interface building tools. The author provides provocative discussion of speech input/output, natural-language interaction, anthropomorphic design, virtual environments, and intelligent (software) agents.
series other
last changed 2003/04/23 15:14

_id avocaad_2001_19
id avocaad_2001_19
authors Shen-Kai Tang, Yu-Tung Liu, Yu-Sheng Chung, Chi-Seng Chung
year 2001
title The visual harmony between new and old materials in the restoration of historical architecture: A study of computer simulation
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In the research of historical architecture restoration, scholars respectively focus on the field of architectural context and architectural archeology (Shi, 1988, 1990, 1991, 1992, 1995; Fu, 1995, 1997; Chiu, 2000) or on architecture construction and the procedure of restoration (Shi, 1988, 1989; Chiu, 1990). How to choose materials and cope with their durability becomes an important issue in the restoration of historical architecture (Dasser, 1990; Wang, 1998).In the related research of the usage and durability of materials, some scholars deem that, instead of continuing the traditional ways that last for hundreds of years (that is to replace new materials with old ones), it might be better to keep the original materials (Dasser, 1990). However, unavoidably, some of the originals are much worn. Thus we have to first establish the standard of eliminating components, and secondly to replace identical or similar materials with the old components (Lee, 1990). After accomplishing the restoration, we often unexpectedly find out that the renewed historical building is too new that the sense of history is eliminated (Dasser, 1990; Fu, 1997). Actually this is the important factor that determines the accomplishment of restoration. In the past, some scholars find out that the contrast and conflict between new and old materials are contributed to the different time of manufacture and different coating, such as antiseptic, pattern, etc., which result in the discrepancy of the sense of visual perception (Lee, 1990; Fu, 1997; Dasser, 1990).In recent years, a number of researches and practice of computer technology have been done in the field of architectural design. We are able to proceed design communication more exactly by the application of some systematic softwares, such as image processing, computer graphic, computer modeling/rendering, animation, multimedia, virtual reality and so on (Lawson, 1995; Liu, 1996). The application of computer technology to the research of the preservation of historical architecture is comparatively late. Continually some researchers explore the procedure of restoration by computer simulation technology (Potier, 2000), or establish digital database of the investigation of historical architecture (Sasada, 2000; Wang, 1998). How to choose materials by the technology of computer simulation influences the sense of visual perception. Liu (2000) has a more complete result on visual impact analysis and assessment (VIAA) about the research of urban design projection. The main subjects of this research paper focuses on whether the technology of computer simulation can extenuate the conflict between new and old materials that imposed on visual perception.The objective of this paper is to propose a standard method of visual harmony effects for materials in historical architecture (taking the Gigi Train Station destroyed by the earthquake in last September as the operating example).There are five steps in this research: 1.Categorize the materials of historical architecture and establish the information in digital database. 2.Get new materials of historical architecture and establish the information in digital database. 3.According to the mixing amount of new and old materials, determinate their proportion of the building; mixing new and old materials in a certain way. 4.Assign the mixed materials to the computer model and proceed the simulation of lighting. 5.Make experts and the citizens to evaluate the accomplished computer model in order to propose the expected standard method.According to the experiment mentioned above, we first address a procedure of material simulation of the historical architecture restoration and then offer some suggestions of how to mix new and old materials.By this procedure of simulation, we offer a better view to control the restoration of historical architecture. And, the discrepancy and discordance by new and old materials can be released. Moreover, we thus avoid to reconstructing ¡§too new¡¨ historical architecture.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id avocaad_2001_20
id avocaad_2001_20
authors Shen-Kai Tang
year 2001
title Toward a procedure of computer simulation in the restoration of historical architecture
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In the field of architectural design, “visualization¨ generally refers to some media, communicating and representing the idea of designers, such as ordinary drafts, maps, perspectives, photos and physical models, etc. (Rahman, 1992; Susan, 2000). The main reason why we adopt visualization is that it enables us to understand clearly and to control complicated procedures (Gombrich, 1990). Secondly, the way we get design knowledge is more from the published visualized images and less from personal experiences (Evans, 1989). Thus the importance of the representation of visualization is manifested.Due to the developments of computer technology in recent years, various computer aided design system are invented and used in a great amount, such as image processing, computer graphic, computer modeling/rendering, animation, multimedia, virtual reality and collaboration, etc. (Lawson, 1995; Liu, 1996). The conventional media are greatly replaced by computer media, and the visualization is further brought into the computerized stage. The procedure of visual impact analysis and assessment (VIAA), addressed by Rahman (1992), is renewed and amended for the intervention of computer (Liu, 2000). Based on the procedures above, a great amount of applied researches are proceeded. Therefore it is evident that the computer visualization is helpful to the discussion and evaluation during the design process (Hall, 1988, 1990, 1992, 1995, 1996, 1997, 1998; Liu, 1997; Sasada, 1986, 1988, 1990, 1993, 1997, 1998). In addition to the process of architectural design, the computer visualization is also applied to the subject of construction, which is repeatedly amended and corrected by the images of computer simulation (Liu, 2000). Potier (2000) probes into the contextual research and restoration of historical architecture by the technology of computer simulation before the practical restoration is constructed. In this way he established a communicative mode among archeologists, architects via computer media.In the research of restoration and preservation of historical architecture in Taiwan, many scholars have been devoted into the studies of historical contextual criticism (Shi, 1988, 1990, 1991, 1992, 1995; Fu, 1995, 1997; Chiu, 2000). Clues that accompany the historical contextual criticism (such as oral information, writings, photographs, pictures, etc.) help to explore the construction and the procedure of restoration (Hung, 1995), and serve as an aid to the studies of the usage and durability of the materials in the restoration of historical architecture (Dasser, 1990; Wang, 1998). Many clues are lost, because historical architecture is often age-old (Hung, 1995). Under the circumstance, restoration of historical architecture can only be proceeded by restricted pictures, written data and oral information (Shi, 1989). Therefore, computer simulation is employed by scholars to simulate the condition of historical architecture with restricted information after restoration (Potier, 2000). Yet this is only the early stage of computer-aid restoration. The focus of the paper aims at exploring that whether visual simulation of computer can help to investigate the practice of restoration and the estimation and evaluation after restoration.By exploring the restoration of historical architecture (taking the Gigi Train Station destroyed by the earthquake in last September as the operating example), this study aims to establish a complete work on computer visualization, including the concept of restoration, the practice of restoration, and the estimation and evaluation of restoration.This research is to simulate the process of restoration by computer simulation based on visualized media (restricted pictures, restricted written data and restricted oral information) and the specialized experience of historical architects (Potier, 2000). During the process of practicing, communicates with craftsmen repeatedly with some simulated alternatives, and makes the result as the foundation of evaluating and adjusting the simulating process and outcome. In this way we address a suitable and complete process of computer visualization for historical architecture.The significance of this paper is that we are able to control every detail more exactly, and then prevent possible problems during the process of restoration of historical architecture.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id fd02
authors Tsou, Jin-Yeu
year 1992
title Using conceptual modelling and an object-oriented environment to support building cost control during early design
source College of Architecture and Urban Planning, University of Michigan
summary This research investigated formal information modelling techniques and the object-oriented knowledge representation on the domain of building cost control during early design stages. The findings contribute to an understanding of the advantages and disadvantages of applying formal modelling techniques to the analysis of architectural problems and the representation of domain knowledge in an object-oriented environment. In this study, information modelling techniques were reviewed, formal information analysis was performed, a conceptual model based on the cost control problem domain was created, a computational model based on the object-oriented approach was developed, a mechanism to support information broadcasting for representing interrelationships was implemented, and an object-oriented cost analysis system for early design (OBCIS) was demonstrated. The conceptual model, based on the elemental proposition analysis of NIAM, supports a formal approach for analyzing the problem domain; the analysis results are represented by high-level graphical notations, based on the AEC Building System Model, to visually display the information framework of the domain. The conceptual model provides an intermediate step between the system designer's view of the domain and the internal representation of the implementation platform. The object-oriented representation provides extensive data modelling abilities to help system designers intuitively represent the semantics of the problem domain. The object-oriented representation also supports more structured and integrated modules than conventional programming approaches. Although there are many advantages to applying this technique to represent the semantics of cost control knowledge, there are several issues which need to be considered: no single satisfactory classification method can be directly applied; object-oriented systems are difficult to learn; and designing reusable classes is difficult. The dependency graph and information broadcasting implemented in this research is an attempt to represent the interrelationships between domain objects. The mechanism allows users to explicitly define the interrelationships, based on semantic requirements, among domain objects. In the conventional approach, these relationships are directly interpreted by system designers and intertwined into the programming code. There are several issues which need to be studied further: indirect dependency relationship, conflict resolution, and request-update looping based on least-commitment approach.
series thesis:PhD
email
last changed 2003/02/12 22:37

_id 60e7
authors Bailey, Rohan
year 2000
title The Intelligent Sketch: Developing a Conceptual Model for a Digital Design Assistant
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 137-145
doi https://doi.org/10.52842/conf.acadia.2000.137
summary The computer is a relatively new tool in the practice of Architecture. Since its introduction, there has been a desire amongst designers to use this new tool quite early in the design process. However, contrary to this desire, most Architects today use pen and paper in the very early stages of design to sketch. Architects solve problems by thinking visually. One of the most important tools that the Architect has at his disposal in the design process is the hand sketch. This iterative way of testing ideas and informing the design process with images fundamentally directs and aids the architect’s decision making. It has been said (Schön and Wiggins 1992) that sketching is about the reflective conversation designers have with images and ideas conveyed by the act of drawing. It is highly dependent on feedback. This “conversation” is an area worthy of investigation. Understanding this “conversation” is significant to understanding how we might apply the computer to enhance the designer’s ability to capture, manipulate and reflect on ideas during conceptual design. This paper discusses sketching and its relation to design thinking. It explores the conversations that designers engage in with the media they use. This is done through the explanation of a protocol analysis method. Protocol analysis used in the field of psychology, has been used extensively by Eastman et al (starting in the early 70s) as a method to elicit information about design thinking. In the pilot experiment described in this paper, two persons are used. One plays the role of the “hand” while the other is the “mind”- the two elements that are involved in the design “conversation”. This variation on classical protocol analysis sets out to discover how “intelligent” the hand should be to enhance design by reflection. The paper describes the procedures entailed in the pilot experiment and the resulting data. The paper then concludes by discussing future intentions for research and the far reaching possibilities for use of the computer in architectural studio teaching (as teaching aids) as well as a digital design assistant in conceptual design.
keywords CAAD, Sketching, Protocol Analysis, Design Thinking, Design Education
series ACADIA
last changed 2022/06/07 07:54

_id e412
authors Fargas, Josep and Papazian, Pegor
year 1992
title Modeling Regulations and Intentions for Urban Development: The Role of Computer Simulation in the Urban Design Studio
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 201-212
doi https://doi.org/10.52842/conf.ecaade.1992.201
summary In this paper we present a strategy for modeling urban development in order to study the role of urban regulations and policies in the transformation of cities. We also suggest a methodology for using computer models as experimental tools in the urban design studio in order to make explicit the factors involved in shaping cities, and for the automatic visualization of projected development. The structure of the proposed model is based on different modules which represent, on the one hand, the rules regulating the physical growth of a city and, on the other hand, heuristics corresponding to different interests such as Real Estate Developers, City Hall Planners, Advocacy and Community Groups, and so on. Here we present a case study dealing with the Boston Redevelopment Authority zoning code for the Midtown Cultural District of Boston. We introduce a computer program which develops the district, adopting a particular point of view regarding urban regulation. We then generalize the notion of this type of computer modeling and simulation, and draw some conclusions about its possible uses in the teaching and practice of design.
series eCAADe
email
last changed 2022/06/07 07:55

_id ddss9207
id ddss9207
authors Gauchel, J., Hovestadt, L., van Wyk, S. and Bhat, R.R.
year 1993
title Modular building models
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary The development and implementation of a modular building model appropriate for computer aided design is described. The limitations of a unified building model with regard to concurrence and complexity in design is discussed. Current research suggests that to model real-world complexity, one must trade centralized control for autonomy. In this paper we develop a modular approach to building modelling that is based on object-oriented autonomy and makes it possible to define these models in a distributed concurrent manner. Such a modular and autonomous implementation brings inherent uncertainty and conflict which cannot be determined a priori.
series DDSS
last changed 2003/08/07 16:36

_id 7e68
authors Holland, J.
year 1992
title Genetic Algorithms
source Scientific America, July 1992
summary Living organisms are consummate problem solvers. They exhibit a versatility that puts the best computer programs to shame. This observation is especially galling for computer scientists, who may spend months or years of intellectual effort on an algorithm, whereas organisms come by their abilities through the apparently undirected mechanism of evolution and natural selection. Pragmatic researchers see evolution's remarkable power as something to be emulated rather than envied. Natural selection eliminates one of the greatest hurdles in software design: specifying in advance all the features of a problem and the actions a program should take to deal with them. By harnessing the mechanisms of evolution, researchers may be able to "breed" programs that solve problems even when no person can fully understand their structure. Indeed, these so-called genetic algorithms have already demonstrated the ability to made breakthroughs in the design of such complex systems as jet engines. Genetic algorithms make it possible to explore a far greater range of potential solutions to a problem than do conventional programs. Furthermore, as researchers probe the natural selection of programs under controlled an well-understood conditions, the practical results they achieve may yield some insight into the details of how life and intelligence evolve in the natural world.
series journal paper
last changed 2003/04/23 15:50

_id 4b2a
id 4b2a
authors Jabi, Wassim
year 2004
title A FRAMEWORK FOR COMPUTER-SUPPORTED COLLABORATION IN ARCHITECTURAL DESIGN
source University of Michigan
summary The development of appropriate research frameworks and guidelines for the construction of software aids in the area of architectural design can lead to a better understanding of designing and computer support for designing (Gero and Maher 1997). The field of research and development in computer-supported collaborative architectural design reflects that of the early period in the development of the field of computersupported cooperative work (CSCW). In the early 1990s, the field of CSCW relied on unsystematic attempts to generate software that increases the productivity of people working together (Robinson 1992). Furthermore, a shift is taking place by which researchers in the field of architecture are increasingly becoming consumers of rather than innovators of technology (Gero and Maher . In particular, the field of architecture is rapidly becoming dependent on commercial software implementations that are slow to respond to new research or to user demands. Additionally, these commercial systems force a particular view of the domain they serve and as such might hinder rather than help its development. The aim of this dissertation is to provide information to architects and others to help them build their own tools or, at a minimum, be critical of commercial solutions.
series thesis:PhD
type normal paper
email
last changed 2004/10/24 22:35

_id ed78
authors Jog, Bharati
year 1993
title Integration of Computer Applications in the Practice of Architecture
source Education and Practice: The Critical Interface [ACADIA Conference Proceedings / ISBN 1-880250-02-0] Texas (Texas / USA) 1993, pp. 89-97
doi https://doi.org/10.52842/conf.acadia.1993.089
summary Computer Applications in Architecture is emerging as an important aspect of our profession. The field, which is often referred to as Computer-Aided Architectural Design (CAAD) has had a notable impact on the profession and academia in recent years. A few professionals have predicted that as slide rules were replaced by calculators, in the coming years drafting boards and parallel bars will be replaced by computers. On the other hand, many architects do not anticipate such a drastic change in the coming decade as present CAD systems are supporting only a few integral aspects of architectural design. However, all agree that architecture curricula should be modified to integrate CAAD education.

In 1992-93, in the Department of Architecture of the 'School of Architecture and interior Design' at the University of Cincinnati, a curriculum committee was formed to review and modify the entire architecture curriculum. Since our profession and academia relate directly to each other, the author felt that while revising the curriculum, the committee should have factual information about CAD usage in the industry. Three ways to obtain such information were thought of, namely (1) conducting person to person or telephone interviews with the practitioners (2) requesting firms to give open- ended feed back and (3) surveying firms by sending a questionnaire. Of these three, the most effective, efficient and suitable method to obtain such information was an organized survey through a questionnaire. In mid December 1992, a survey was organized which was sponsored by the School of Architecture and Interior Design, the Center for the Study of the Practice of Architecture (CSPA) and the University Division of Professional Practice, all from the University of Cincinnati.

This chapter focuses on the results of this survey. A brief description of the survey design is also given. In the next section a few surveys organized in recent years are listed. In the third section the design of this survey is presented. The survey questions and their responses are given in the fourth section. The last section presents the conclusions and brief recommendations regarding computer curriculum in architecture.

series ACADIA
last changed 2022/06/07 07:52

_id 6d34
authors Kensek, Karen and Noble, Doug (Eds.)
year 1992
title Mission - Method - Madness [Conference Proceedings]
source ACADIA Conference Proceedings / ISBN 1-880250-01-2 ) 1992, 232 p.
doi https://doi.org/10.52842/conf.acadia.1992
summary The papers represent a wide variety of exploration into the uses of computers in architecture. We have tried to impose order onto the collection by organizing them into six sessions: Metaphor, Mission, Method, Modeling for Visualization, Modeling, and Generative Systems. As with any ordering system for such a diverse selection, some session papers are strongly related while others are loosely grouped. Madness, an additional session not in the proceedings, will include short presentations of work in progress. Regarding the individual papers, it is particularly exciting to see research being conducted that is founded on previous work done by others. It is also interesting to note that half of the papers have been submitted by teams of authors. Whether this represents "computer supported cooperative work" remains to be seen. Certainly the work in this book represents an interesting and wide variety of explorations into computer supported design in architecture.
series ACADIA
email
more http://www.acadia.org
last changed 2022/06/07 07:49

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 12HOMELOGIN (you are user _anon_695196 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002