CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 30

_id 735a
authors Anh, Tran Hoai
year 1992
title FULL-SCALE EXPERIMENT ON KITCHEN FUNCTION IN HANOI
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part A, pp. 19-30
summary This study is a part of a licentiate thesis on "Functional kitchen for the Vietnamese cooking way"at the Department of Architecture and Development studies, Lund University. The issues it is dealing with are: (1) Inadequacy of kitchen design in the apartment buildings in Hanoi, where the kitchen is often designed as a mere cooking place - other parts of the food making process are not given any attention. (2) Lack of standard dimensional and planning criteria for functional kitchen which can serve as bases for kitchen design. // The thesis aims at finding out indicators on functional spatial requirements for kitchen, which can serve as guide-line for designing functional kitchen for Hanoi. One of the main propositions in the thesis is that functional kitchens for Hanoi should be organised to permit the culinary activities done according to the Vietnamese urban culinary practice. This is based on the concept that the culinary activity is an expression Of culture, thus the practice of preparing meal in the present context of the urban households in Hanoi has an established pattern, method which demand a suitable area and arrangement in the kitchen. This pattern and cooking method should make up the functional requirement for kitchen in Hanoi, and be taken in to account if functional kitchen designing is to be achieved. In the context of the space-limited apartment building of Hanoi, special focus is given to find out indicators on the minimum functional spatial requirements of the kitchen works.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 15:29

_id 6208
authors Abou-Jaoude, Georges
year 1992
title To Master a Tool
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part B, p. 15
summary The tool here is the computer or to be precise, a unit that includes the computer, the peripherals and the software needed to fulfill a task. These tools are getting very sophisticated and user interfaces extremly friendly, therefore it is very easy to become the slave of such electronic tools and reach self satisfaction with strait forward results and attractive images. In order to master and not to become slaves of sophisticated tools, a very solid knowledge of related fields or domains of application becomes necessary. In the case of this seminar, full scale modelling, is a way to understand the relation between a mental model and it's full-scale modelling, it is a way of communicating what is in a designers mind. Computers and design programs can have the same goal, rather than chosing one method or the other let us try to say how important it is today to complement designing with computer with other means and media such as full scale modelling, and what computer modelling and simulation can bring to full scale modelling or other means.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
more http://info.tuwien.ac.at/efa
last changed 2003/08/25 10:12

_id 60e7
authors Bailey, Rohan
year 2000
title The Intelligent Sketch: Developing a Conceptual Model for a Digital Design Assistant
doi https://doi.org/10.52842/conf.acadia.2000.137
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 137-145
summary The computer is a relatively new tool in the practice of Architecture. Since its introduction, there has been a desire amongst designers to use this new tool quite early in the design process. However, contrary to this desire, most Architects today use pen and paper in the very early stages of design to sketch. Architects solve problems by thinking visually. One of the most important tools that the Architect has at his disposal in the design process is the hand sketch. This iterative way of testing ideas and informing the design process with images fundamentally directs and aids the architect’s decision making. It has been said (Schön and Wiggins 1992) that sketching is about the reflective conversation designers have with images and ideas conveyed by the act of drawing. It is highly dependent on feedback. This “conversation” is an area worthy of investigation. Understanding this “conversation” is significant to understanding how we might apply the computer to enhance the designer’s ability to capture, manipulate and reflect on ideas during conceptual design. This paper discusses sketching and its relation to design thinking. It explores the conversations that designers engage in with the media they use. This is done through the explanation of a protocol analysis method. Protocol analysis used in the field of psychology, has been used extensively by Eastman et al (starting in the early 70s) as a method to elicit information about design thinking. In the pilot experiment described in this paper, two persons are used. One plays the role of the “hand” while the other is the “mind”- the two elements that are involved in the design “conversation”. This variation on classical protocol analysis sets out to discover how “intelligent” the hand should be to enhance design by reflection. The paper describes the procedures entailed in the pilot experiment and the resulting data. The paper then concludes by discussing future intentions for research and the far reaching possibilities for use of the computer in architectural studio teaching (as teaching aids) as well as a digital design assistant in conceptual design.
keywords CAAD, Sketching, Protocol Analysis, Design Thinking, Design Education
series ACADIA
last changed 2022/06/07 07:54

_id e039
authors Bertin, Vito
year 1992
title Structural Transformations (Basic Architectural Unit 6)
doi https://doi.org/10.52842/conf.ecaade.1992.413
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 413-426
summary While the teaching of the phenomenon of form as well as space is normally seen within an environment of free experimentation and personal expression, other directions prove to be worth of pursuit. The proposed paper represents such an exploration. The generation of controlled complexity and structural transformations have been the title of the project which forms the base of this paper. In it, the potential for creative development of the student was explored in such a way, that as in the sciences a process can be reproduced or an exploration utilized in further experimentation. The cube as a well proven B.A.U. or basic architectural unit has again been used in our work. Even a simple object like a cube has many properties. As properties are never pure, but always related to other properties, and looking at a single property as a specific value of a variable, it is possible to link a whole field of objects. These links provide a network of paths through which exploration and development is possible. The paper represents a first step in a direction which we think will compliment the already established basic design program.

series eCAADe
email
last changed 2022/06/07 07:52

_id eabb
authors Boeykens, St. Geebelen, B. and Neuckermans, H.
year 2002
title Design phase transitions in object-oriented modeling of architecture
doi https://doi.org/10.52842/conf.ecaade.2002.310
source Connecting the Real and the Virtual - design e-ducation [20th eCAADe Conference Proceedings / ISBN 0-9541183-0-8] Warsaw (Poland) 18-20 September 2002, pp. 310-313
summary The project IDEA+ aims to develop an “Integrated Design Environment for Architecture”. Its goal is providing a tool for the designer-architect that can be of assistance in the early-design phases. It should provide the possibility to perform tests (like heat or cost calculations) and simple simulations in the different (early) design phases, without the need for a fully detailed design or remodeling in a different application. The test for daylighting is already in development (Geebelen, to be published). The conceptual foundation for this design environment has been laid out in a scheme in which different design phases and scales are defined, together with appropriate tests at the different levels (Neuckermans, 1992). It is a translation of the “designerly” way of thinking of the architect (Cross, 1982). This conceptual model has been translated into a “Core Object Model” (Hendricx, 2000), which defines a structured object model to describe the necessary building model. These developments form the theoretical basis for the implementation of IDEA+ (both the data structure & prototype software), which is currently in progress. The research project addresses some issues, which are at the forefront of the architect’s interest while designing with CAAD. These are treated from the point of view of a practicing architect.
series eCAADe
email
last changed 2022/06/07 07:52

_id 9b34
authors Butterworth, J. (et al.)
year 1992
title 3DM: A three-dimensional modeler using a head-mounted display
source Proceedings of the 1992 Symposium on Interactive 3D Graphics (Cambridge, Mass., March 29- April 1, 1992.), 135-138
summary 3dm is a three dimensional (3D) surface modeling program that draws techniques of model manipulation from both CAD and drawing programs and applies them to modeling in an intuitive way. 3dm uses a head-mounted display (HMD) to simplify the problem of 3D model manipulation and understanding. A HMD places the user in the modeling space, making three dimensional relationships more understandable. As a result, 3dm is easy to learn how to use and encourages experimentation with model shapes.
series other
last changed 2003/04/23 15:50

_id c434
authors Colajanni, B., Pellitteri, G. and Scianna, A.
year 1992
title Two Approaches to Teaching Computers in Architecture: The Experience in the Faculty of Engineering in Palermo, Italy
doi https://doi.org/10.52842/conf.ecaade.1992.295
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 295-306
summary Teaching the use of computers in architecture poses the same kind of problems as teaching mathematics. To both there are two possible approaches. The first presents the discipline as a tool of which the merely instrumental aspect is emphasized. Teaching is limited to show the results obtainable by existing programs and how to get them. The second approach, on the contrary emphasizes the autonomous nature of the discipline, mathematics as much as computing, on the basis of the convincement that the maximum of instrumental usefulness can be obtained through the knowledge at the highest degree of generality and, then, of abstraction. The first approach changes little in the mind of the student. He simply learns that is possible, and then worthy doing, a certain amount of operations, mainly checks of performances (and not only the control of the aspect, now easy with one of the many existing CAD) or searches of technical informations in some database. The second approach gives the student the consciousness of the manageability of abstract structures of relationships. He acquires then the idea of creating by himself particular structures of relationships and managing them. This can modify the very idea of the design procedure giving the student the consciousness that he can intervene directly in every segment of the design procedure, reshaping it to some extent in a way better suited to the particular problem he is dealing with. Of course this second approach implies learning not only a language but also the capability of coming to terms with languages. And again it is a cultural acquisition that can be very useful when referred to the languages of architecture. Furthermore the capability of simulating on the computer also a small segment of the design process gives the student a better understanding both of the particular problem he is dealing with and of the very nature of design. As for the first effect, it happens whenever a translation is done from a language to another one. One is obliged to get to the core of the matter in order to overcome the difficulties rising from the different bias of the two languages. The second effect comes from the necessity of placing the studied segment in the general flow of the design process. The organisation in a linear sequence of action to be accomplished recursively in an order always varying in any design occasion is an extremely useful exercise to understand the signification and the techniques of formalisation of design problems.
series eCAADe
email
last changed 2022/06/07 07:56

_id 4857
authors Escola Tecnica Superior D'arquitectura de Barcelona (Ed.)
year 1992
title CAAD Instruction: The New Teaching of an Architect?
doi https://doi.org/10.52842/conf.ecaade.1992
source eCAADe Conference Proceedings / Barcelona (Spain) 12-14 November 1992, 551 p.
summary The involvement of computer graphic systems in the transmission of knowledge in the areas of urban planning and architectural design will bring a significant change to the didactic programs and methods of those schools which have decided to adopt these new instruments. Workshops of urban planning and architectural design will have to modify their structures, and teaching teams will have to revise their current programs. Some european schools and faculties of architecture have taken steps in this direction. Others are willing to join them.

This process is only delayed by the scarcity of material resources, and by the slowness with which a sufficient number of teachers are adopting these methods.

ECAADE has set out to analyze the state of this issue during its next conference, and it will be discussed from various points of view. From this confrontation of ideas will come, surely, the guidelines for progress in the years to come.

The different sessions will be grouped together following these four themes:

(A.) Multimedia and Course Work / State of the art of the synthesis of graphical and textual information favored by new available multimedia computer programs. Their repercussions on academic programs. (B.) The New Design Studio / Physical characteristics, data concentration and accessibility of a computerized studio can be better approached in a computerized workshop. (C.) How to manage the new education system / Problems and possibilities raised, from the practical and organizational points of view, of architectural education by the introduction of computers in the classrooms. (D.) CAAI. Formal versus informal structure / How will the traditional teaching structure be affected by the incidence of these new systems in which the access to knowledge and information can be obtained in a random way and guided by personal and subjective criteria.

series eCAADe
email
last changed 2022/06/07 07:49

_id 4129
authors Fargas, Josep and Papazian, Pegor
year 1992
title Metaphors in Design: An Experiment with a Frame, Two Lines and Two Rectangles
doi https://doi.org/10.52842/conf.acadia.1992.013
source Mission - Method - Madness [ACADIA Conference Proceedings / ISBN 1-880250-01-2] 1992, pp. 13-22
summary The research we will discuss below originated from an attempt to examine the capacity of designers to evaluate an artifact, and to study the feasibility of replicating a designer's moves intended to make an artifact more expressive of a given quality. We will present the results of an interactive computer experiment, first developed at the MIT Design Research Seminar, which is meant to capture the subject’s actions in a simple design task as a series of successive "moves"'. We will propose that designers use metaphors in their interaction with design artifacts and we will argue that the concept of metaphors can lead to a powerful theory of design activity. Finally, we will show how such a theory can drive the project of building a design system.

When trying to understand how designers work, it is tempting to examine design products in order to come up with the principles or norms behind them. The problem with such an approach is that it may lead to a purely syntactical analysis of design artifacts, failing to capture the knowledge of the designer in an explicit way, and ignoring the interaction between the designer and the evolving design. We will present a theory about design activity based on the observation that knowledge is brought into play during a design task by a process of interpretation of the design document. By treating an evolving design in terms of the meanings and rules proper to a given way of seeing, a designer can reduce the complexity of a task by focusing on certain of its aspects, and can manipulate abstract elements in a meaningful way.

series ACADIA
email
last changed 2022/06/07 07:55

_id 7ce5
authors Gal, Shahaf
year 1992
title Computers and Design Activities: Their Mediating Role in Engineering Education
source Sociomedia, ed. Edward Barret. MIT Press
summary Sociomedia: With all the new words used to describe electronic communication (multimedia, hypertext, cyberspace, etc.), do we need another one? Edward Barrett thinks we do; hence, he coins the term "sociomedia." It is meant to displace a computing economy in which technicity is hypostasized over sociality. Sociomedia, a compilation of twenty-five articles on the theory, design and practice of educational multimedia and hypermedia, attempts to re-value the communicational face of computing. Value, of course, is "ultimately a social construct." As such, it has everything to do with knowledge, power, education and technology. The projects discussed in this book represent the leading edge of electronic knowledge production in academia (not to mention major funding) and are determining the future of educational media. For these reasons, Sociomedia warrants close inspection. Barrett's introduction sets the tone. For him, designing computer media involves hardwiring a mechanism for the social construction of knowledge (1). He links computing to a process of social and communicative interactivity for constructing and desseminating knowledge. Through a mechanistic mapping of the university as hypercontext (a huge network that includes classrooms as well as services and offices), Barrett models intellectual work in such a way as to avoid "limiting definitions of human nature or human development." Education, then, can remain "where it should be--in the human domain (public and private) of sharing ideas and information through the medium of language." By leaving education in a virtual realm (where we can continue to disagree about its meaning and execution), it remains viral, mutating and contaminating in an intellectually healthy way. He concludes that his mechanistic model, by means of its reductionist approach, preserves value (7). This "value" is the social construction of knowledge. While I support the social orientation of Barrett's argument, discussions of value are related to power. I am not referring to the traditional teacher-student power structure that is supposedly dismantled through cooperative and constructivist learning strategies. The power to be reckoned with in the educational arena is foundational, that which (pre)determines value and the circulation of knowledge. "Since each of you reading this paragraph has a different perspective on the meaning of 'education' or 'learning,' and on the processes involved in 'getting an education,' think of the hybris in trying to capture education in a programmable function, in a displayable object, in a 'teaching machine'" (7). Actually, we must think about that hybris because it is, precisely, what informs teaching machines. Moreover, the basic epistemological premises that give rise to such productions are too often assumed. In the case of instructional design, the episteme of cognitive sciences are often taken for granted. It is ironic that many of the "postmodernists" who support electronic hypertextuality seem to have missed Jacques Derrida's and Michel Foucault's "deconstructions" of the epistemology underpinning cognitive sciences (if not of epistemology itself). Perhaps it is the glitz of the technology that blinds some users (qua developers) to the belief systems operating beneath the surface. Barrett is not guilty of reactionary thinking or politics; he is, in fact, quite in line with much American deconstructive and postmodern thinking. The problem arises in that he leaves open the definitions of "education," "learning" and "getting an education." One cannot engage in the production of new knowledge without orienting its design, production and dissemination, and without negotiating with others' orientations, especially where largescale funding is involved. Notions of human nature and development are structural, even infrastructural, whatever the medium of the teaching machine. Although he addresses some dynamics of power, money and politics when he talks about the recession and its effects on the conference, they are readily visible dynamics of power (3-4). Where does the critical factor of value determination, of power, of who gets what and why, get mapped onto a mechanistic model of learning institutions? Perhaps a mapping of contributors' institutions, of the funding sources for the projects showcased and for participation in the conference, and of the disciplines receiving funding for these sorts of projects would help visualize the configurations of power operative in the rising field of educational multimedia. Questions of power and money notwithstanding, Barrett's introduction sets the social and textual thematics for the collection of essays. His stress on interactivity, on communal knowledge production, on the society of texts, and on media producers and users is carried foward through the other essays, two of which I will discuss. Section I of the book, "Perspectives...," highlights the foundations, uses and possible consequences of multimedia and hypertextuality. The second essay in this section, "Is There a Class in This Text?," plays on the robust exchange surrounding Stanley Fish's book, Is There a Text in This Class?, which presents an attack on authority in reading. The author, John Slatin, has introduced electronic hypertextuality and interaction into his courses. His article maps the transformations in "the content and nature of work, and the workplace itself"-- which, in this case, is not industry but an English poetry class (25). Slatin discovered an increase of productive and cooperative learning in his electronically- mediated classroom. For him, creating knowledge in the electronic classroom involves interaction between students, instructors and course materials through the medium of interactive written discourse. These interactions lead to a new and persistent understanding of the course materials and of the participants' relation to the materials and to one another. The work of the course is to build relationships that, in my view, constitute not only the meaning of individual poems, but poetry itself. The class carries out its work in the continual and usually interactive production of text (31). While I applaud his strategies which dismantle traditional hierarchical structures in academia, the evidence does not convince me that the students know enough to ask important questions or to form a self-directing, learning community. Stanley Fish has not relinquished professing, though he, too, espouses the indeterminancy of the sign. By the fourth week of his course, Slatin's input is, by his own reckoning, reduced to 4% (39). In the transcript of the "controversial" Week 6 exchange on Gertrude Stein--the most disliked poet they were discussing at the time (40)--we see the blind leading the blind. One student parodies Stein for three lines and sums up his input with "I like it." Another, finds Stein's poetry "almost completey [sic] lacking in emotion or any artistic merit" (emphasis added). On what grounds has this student become an arbiter of "artistic merit"? Another student, after admitting being "lost" during the Wallace Steven discussion, talks of having more "respect for Stevens' work than Stein's" and adds that Stein's poetry lacks "conceptual significance[, s]omething which people of varied opinion can intelligently discuss without feeling like total dimwits...." This student has progressed from admitted incomprehension of Stevens' work to imposing her (groundless) respect for his work over Stein's. Then, she exposes her real dislike for Stein's poetry: that she (the student) missed the "conceptual significance" and hence cannot, being a person "of varied opinion," intelligently discuss it "without feeling like [a] total dimwit." Slatin's comment is frightening: "...by this point in the semester students have come to feel increasingly free to challenge the instructor" (41). The students that I have cited are neither thinking critically nor are their preconceptions challenged by student-governed interaction. Thanks to the class format, one student feels self-righteous in her ignorance, and empowered to censure. I believe strongly in student empowerment in the classroom, but only once students have accrued enough knowledge to make informed judgments. Admittedly, Slatin's essay presents only partial data (there are six hundred pages of course transcripts!); still, I wonder how much valuable knowledge and metaknowledge was gained by the students. I also question the extent to which authority and professorial dictature were addressed in this course format. The power structures that make it possible for a college to require such a course, and the choice of texts and pedagogy, were not "on the table." The traditional professorial position may have been displaced, but what took its place?--the authority of consensus with its unidentifiable strong arm, and the faceless reign of software design? Despite Slatin's claim that the students learned about the learning process, there is no evidence (in the article) that the students considered where their attitudes came from, how consensus operates in the construction of knowledge, how power is established and what relationship they have to bureaucratic insitutions. How do we, as teaching professionals, negotiate a balance between an enlightened despotism in education and student-created knowledge? Slatin, and other authors in this book, bring this fundamental question to the fore. There is no definitive answer because the factors involved are ultimately social, and hence, always shifting and reconfiguring. Slatin ends his article with the caveat that computerization can bring about greater estrangement between students, faculty and administration through greater regimentation and control. Of course, it can also "distribute authority and power more widely" (50). Power or authority without a specific face, however, is not necessarily good or just. Shahaf Gal's "Computers and Design Activities: Their Mediating Role in Engineering Education" is found in the second half of the volume, and does not allow for a theory/praxis dichotomy. Gal recounts a brief history of engineering education up to the introduction of Growltiger (GT), a computer-assisted learning aid for design. He demonstrates GT's potential to impact the learning of engineering design by tracking its use by four students in a bridge-building contest. What his text demonstrates clearly is that computers are "inscribing and imaging devices" that add another viewpoint to an on-going dialogue between student, teacher, earlier coursework, and other teaching/learning tools. The less proficient students made a serious error by relying too heavily on the technology, or treating it as a "blueprint provider." They "interacted with GT in a way that trusted the data to represent reality. They did not see their interaction with GT as a negotiation between two knowledge systems" (495). Students who were more thoroughly informed in engineering discourses knew to use the technology as one voice among others--they knew enough not simply to accept the input of the computer as authoritative. The less-advanced students learned a valuable lesson from the competition itself: the fact that their designs were not able to hold up under pressure (literally) brought the fact of their insufficient knowledge crashing down on them (and their bridges). They also had, post factum, several other designs to study, especially the winning one. Although competition and comparison are not good pedagogical strategies for everyone (in this case the competitors had volunteered), at some point what we think we know has to be challenged within the society of discourses to which it belongs. Students need critique in order to learn to push their learning into auto-critique. This is what is lacking in Slatin's discussion and in the writings of other avatars of constructivist, collaborative and computer-mediated pedagogies. Obviously there are differences between instrumental types of knowledge acquisition and discoursive knowledge accumulation. Indeed, I do not promote the teaching of reading, thinking and writing as "skills" per se (then again, Gal's teaching of design is quite discursive, if not dialogic). Nevertheless, the "soft" sciences might benefit from "bridge-building" competitions or the re-institution of some forms of agonia. Not everything agonistic is inhuman agony--the joy of confronting or creating a sound argument supported by defensible evidence, for example. Students need to know that soundbites are not sound arguments despite predictions that electronic writing will be aphoristic rather than periodic. Just because writing and learning can be conceived of hypertextually does not mean that rigor goes the way of the dinosaur. Rigor and hypertextuality are not mutually incompatible. Nor is rigorous thinking and hard intellectual work unpleasurable, although American anti-intellectualism, especially in the mass media, would make it so. At a time when the spurious dogmatics of a Rush Limbaugh and Holocaust revisionist historians circulate "aphoristically" in cyberspace, and at a time when knowledge is becoming increasingly textualized, the role of critical thinking in education will ultimately determine the value(s) of socially constructed knowledge. This volume affords the reader an opportunity to reconsider knowledge, power, and new communications technologies with respect to social dynamics and power relationships.
series other
last changed 2003/04/23 15:14

_id 32eb
authors Henry, Daniel
year 1992
title Spatial Perception in Virtual Environments : Evaluating an Architectural Application
source University of Washington
summary Over the last several years, professionals from many different fields have come to the Human Interface Technology Laboratory (H.I.T.L) to discover and learn about virtual environments. In general, they are impressed by their experiences and express the tremendous potential the tool has in their respective fields. But the potentials are always projected far in the future, and the tool remains just a concept. This is justifiable because the quality of the visual experience is so much less than what people are used to seeing; high definition television, breathtaking special cinematographic effects and photorealistic computer renderings. Instead, the models in virtual environments are very simple looking; they are made of small spaces, filled with simple or abstract looking objects of little color distinctions as seen through displays of noticeably low resolution and at an update rate which leaves much to be desired. Clearly, for most applications, the requirements of precision have not been met yet with virtual interfaces as they exist today. However, there are a few domains where the relatively low level of the technology could be perfectly appropriate. In general, these are applications which require that the information be presented in symbolic or representational form. Having studied architecture, I knew that there are moments during the early part of the design process when conceptual decisions are made which require precisely the simple and representative nature available in existing virtual environments. This was a marvelous discovery for me because I had found a viable use for virtual environments which could be immediately beneficial to architecture, my shared area of interest. It would be further beneficial to architecture in that the virtual interface equipment I would be evaluating at the H.I.T.L. happens to be relatively less expensive and more practical than other configurations such as the "Walkthrough" at the University of North Carolina. The set-up at the H.I.T.L. could be easily introduced into architectural firms because it takes up very little physical room (150 square feet) and it does not require expensive and space taking hardware devices (such as the treadmill device for simulating walking). Now that the potential for using virtual environments in this architectural application is clear, it becomes important to verify that this tool succeeds in accurately representing space as intended. The purpose of this study is to verify that the perception of spaces is the same, in both simulated and real environment. It is hoped that the findings of this study will guide and accelerate the process by which the technology makes its way into the field of architecture.
keywords Space Perception; Space (Architecture); Computer Simulation
series thesis:MSc
last changed 2003/02/12 22:37

_id 6e99
authors Hoffer, Erin Rae
year 1992
title Creating the Electronic Design Studio: Development of a Heterogeneous Networked Environment at Harvard's Graduate School of Design
doi https://doi.org/10.52842/conf.ecaade.1992.225
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 225-240
summary The migration of design education to reliance on computer-based techniques requires new ways of thinking about environments which can effectively support a diverse set of activities. Both from a spatial standpoint and a computing resource standpoint, design studios must be inevitably reconfigured to support new tools and reflect new ways of communicating. At Harvard's GSD, a commitment to incorporating computer literacy as a fundamental component of design education enables us to confront these issues through the implementation of a heterogeneous network imbedded in an electronic design environment. This evolving prototype of a new design studio, its development and its potential, will be the subject of this paper. A new style design environment is built upon an understanding of traditional techniques, and layered with an awareness of new tools and methods. Initially we borrow from existing metaphors which govern our interpretation of the way designers work. Next we seek to extend our thinking to include allied or related metaphors such as the library metaphor which informs collections of software and data, or the laboratory metaphor which informs workspace groupings, or the transportation metaphor which informs computer-based communications such as electronic mail or bulletin boards, or the utility services metaphor which informs the provision of network services and equipment. Our evaluation of this environment is based on direct feedback from its users, both faculty and students, and on subjective observation of the qualitative changes in communication which occur between and among these groups and individuals. Ultimately, the network must be judged as a framework for learning and evaluation, and its success depends both on its ability to absorb our existing metaphors for the process of design, and to prefigure the emerging metaphors to be envisioned in the future.

series eCAADe
last changed 2022/06/07 07:50

_id 88ca
authors Kane, Andy and Szalapaj, Peter
year 1992
title Teaching Design By Analysis of Precedents
doi https://doi.org/10.52842/conf.ecaade.1992.477
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 477-496
summary Designers, using their intuitive understanding of the decomposition of particular design objects, whether in terms of structural, functional, or some other analytical framework, should be able to interact with computational environments such that the understanding they achieve in turn invokes changes or transformations to the spatial properties of design proposals. Decompositions and transformations of design precedents can be a very useful method of enabling design students to develop analytical strategies. The benefit of an analytical approach is that it can lead to a structured understanding of design precedents. This in turn allows students to develop their own insights and ideas which are central to the activity of designing. The creation of a 3-D library of user-defined models of precedents in a computational environment permits an under-exploited method of undertaking analysis, since by modelling design precedents through the construction of 3-D Computer-Aided Architectural Design (CAAD) models, and then analytically decomposing them in terms of relevant features, significant insights into the nature of designs can be achieved. Using CAAD systems in this way, therefore, runs counter to the more common approach of detailed modelling, rendering and animation; which produces realistic pictures that do not reflect the design thinking that went into their production. The significance of the analytical approach to design teaching is that it encourages students to represent design ideas, but not necessarily the final form of design objects. The analytical approach therefore, allows students to depict features and execute tasks that are meaningful with respect to design students' own knowledge of particular domains. Such computational interaction can also be useful in helping students explore the consequences of proposed actions in actual design contexts.
series eCAADe
last changed 2022/06/07 07:52

_id ca47
authors Lee, Shu Wan
year 1996
title A Cognitive Approach to Architectural Style Several Characteristics of Design Thinking in Architecture
doi https://doi.org/10.52842/conf.caadria.1996.223
source CAADRIA ‘96 [Proceedings of The First Conference on Computer Aided Architectural Design Research in Asia / ISBN 9627-75-703-9] Hong Kong (Hong Kong) 25-27 April 1996, pp. 223-226
summary Designing is a complicated human behaviour and method, and is often treated as a mysterious "black box” operation in human mind. In the early period as for theory-studying of design thinking, the way of thinking that the researchers took were mostly descriptive discussions. Therefore, they lacked direct and empirical evidence although those studies provided significant exploration of design thinking (Wang, 1995). In recent years as for the study of cognitive science, they have tried to make design "glass box”. That is to try to make the thinking processes embedded in designers publicized. That is also to externalize the design procedure which provided the design studies another theoretical basis of more accurate and deeply researched procedure (Jones, 1992). Hence the studying of design thinking has become more important and the method of designing has also progressed a lot. For example, the classification of the nature of design problem such as ill-defined and well-defined (Newell, Shaw, and Simon, 1967), and different theoretical procedure modes for different disciplines, such as viewing architectural models as conjecture-analysis models and viewing engineering models as analysis-synthesis (Cross, 1991).
series CAADRIA
last changed 2022/06/07 07:52

_id 8b12
authors Manning, Peter and Mattar, Samir
year 1992
title A Preliminary to Development of Expert Systems for Total Design of Entire Buildings
source New York: John Wiley & Sons, 1992. pp. 215-237 : tables. includes bibliography
summary This paper has two primary objectives. The first is to represent the practicability of making the design of entire buildings a conscious, craftsman-like, activity conducted in the clear, without the mystery that tends, because of designers' usual 'black box' methods, to surround it. To this end, a design strategy and some tactics for resolving decisions at critical stages in the design process, which the authors have described elsewhere, are recapitulated to show how total design of buildings can be pursued in a generic manner. This done, the way is opened for the second objective: to make the large and important field of work that is building design amenable to computerization. The form that pursuit of this second objective is taking is being influenced greatly by growing interest in expert systems, which for everyday professional building design appears a more useful development than previous CAD emphases on drafting and graphics. Application of the authors' design methods to a series of expert systems for the total design of entire buildings is therefore indicated. For such a vast project--the formulation of bases for design assistance and expert systems that can be integrated and used as a generic method for the total design of entire buildings, so that the results are more certain and successful than the outcome of the generality of present-day building design--the most that can be attempted within the limits of a single paper is a set of examples of some of the stages in the process. Nevertheless, since the design method described begins at the 'large end' of the process, where the most consequential decisions are made, it is hoped that the major thrusts and the essential CAD activities will be evident. All design is substantially iterative, and provided that the major iterations are intelligible, there should be no need for this demonstration to labor over the lesser ones
keywords evaluation, integration, architecture, building, expert systems, design methods, design process
series CADline
last changed 2003/06/02 13:58

_id 2c22
authors O'Neill, Michael J.
year 1992
title Neural Network Simulation as a Computer- Aided design Tool For Predicting Wayfinding Performance
source New York: John Wiley & Sons, 1992. pp. 347-366 : ill. includes bibliography
summary Complex public facilities such as libraries, hospitals, and governmental buildings often present problems to users who must find their way through them. Research shows that difficulty in wayfinding has costs in terms of time, money, public safety, and stress that results from being lost. While a wide range of architectural research supports the notion that ease of wayfinding should be a criterion for good design, architects have no method for evaluating how well their building designs will support the wayfinding task. People store and retrieve information about the layout of the built environment in a knowledge representation known as the cognitive map. People depend on the information stored in the cognitive map to find their way through buildings. Although there are numerous simulations of the cognitive map, the mechanisms of these models are not constrained by what is known about the neurophysiology of the brain. Rather, these models incorporate search mechanisms that act on semantically encoded information about the environment. In this paper the author describes the evaluation and application of an artificial neural network simulation of the cognitive map as a means of predicting wayfinding behavior in buildings. This simulation is called NAPS-PC (Network Activity Processing Simulator--PC version). This physiologically plausible model represents knowledge about the layout of the environment through a network of inter-connected processing elements. The performance of NAPS-PC was evaluated against actual human wayfinding performance. The study found that the simulation generated behavior that matched the performance of human participants. After the validation, NAPS-PC was modified so that it could read environmental information directly from AutoCAD (a popular micro-computer-based CAD software package) drawing files, and perform 'wayfinding' tasks based on that environmental information. This prototype tool, called AutoNet, is conceptualized as a means of allowing designers to predict the wayfinding performance of users in a building before it is actually built
keywords simulation, cognition, neural networks, evaluation, floor plans, applications, wayfinding, layout, building
series CADline
last changed 2003/06/02 13:58

_id 1850
authors Palmqvist, Henri
year 1992
title SAD - SIMULATOR AIDED DESIGN
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part B, pp. 39-42
summary The multi-level, multidimensional nature of architecture means that the only way to experience it and evaluate it is by moving within the buildings and the environment they constitute. Until today, architectural projects could only be observed after they had been finished. The recently developed environmental simulator is the solution to the problem. With the simulator and scale models we can observe the experience of moving around within the buildings and thus make a significant improvement in the certainty of the assumed architectural experience and content while the project is still under design.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 15:40

_id ddss9210
id ddss9210
authors Poortman, E.R.
year 1993
title Ratios for cost control
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary The design of buildings takes place in phases representing a development from rough to precision planning. Estimates are made in order to test whether the result is still within the budget set by the client or developer. In this way, the decisions taken during the design phase can be quantified and expressed in monetary terms. To prevent blaming the wrong person when an overrun is discovered, the cost control process has to be improved. For that purpose, two new procedures have been developed: (i) a new translation activity; and (ii) ratios by which quantities can be characterized. 'Translation is the opposite of estimation. A monetary budget is converted -'translated' - into quantities, reflecting the desired quality of the building materials. The financial constraints of the client are thus converted into quantities - the building components used by the designers. Characteristic quantity figures play an important role in this activity. In working out an estimate, the form factor (i.e., the ratio between two characteristic values of a building component) has to be determined. The unit cost is then tested against that ratio. The introduction of the 'translation' activity and the use of characteristic quantity figures and form factors enhance existing estimation methods. By implementing these procedures, cost control becomes considerably more reliable.
series DDSS
last changed 2003/08/07 16:36

_id 0c59
authors Redondo, Ernest and Monedero, Javier
year 1992
title Electronic Collage
doi https://doi.org/10.52842/conf.ecaade.1992.241
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 241-250
summary A painting is always a potential form of collage. The discovery made by Braque and Picasso was already implied in the carefully-done executions of tapestries backgrounds and mural decorating of all the italian and flemish paintings which started to be produced since Giotto's, where the represented and the presented were mixed up. And also, in a more pedantic way, in a famous painting of Courbet "L'Atelier du Peintre, allegorie reelle" (1885). In this work, the artist is shown sitting in front of his work, grasping a brush with his right hand while holding a palette, which is at the centre of the composition, in his left hand. There are some figures surrounding the master, some nearer, others further away; in the corners of the atelier one can still see silhouettes which we don't really know whether they belong to real characters or to painted ones on the master's canvases. This ambiguity brings us back to the foreground, to this central palette where the colours, distributed in heaps are, simultaneously, concretion and representation, real matter and figurative matter.

series eCAADe
email
last changed 2022/06/07 08:00

_id a5cc
authors Sabater, Txatxo and Gassull, Albert
year 1992
title From Notion to Motion
doi https://doi.org/10.52842/conf.ecaade.1992.543
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 543-551
summary Going from notion to motion is a way, or a working system. It means the illustration in motion of critical written topics. It's also an indirect channel to normalize the use of CAD and other kind of software and periferials in a School of Architecture held only by a user technology. We deal with texts and the choice of these is absolutely determinant. First of all because of the volition of using those which time has allowed to clearly decant and now are seen together with the answers or continuities that they have generated. That is to say, we do not write on the subjects we talk about, we illustrate, in motion, the arguments that authors have already written about them. We refer to notion in the sense that we always set off from a seminal argument. But also because we collect, if necessary, its revisions or extensions. This is to say we try to track the notion helping ourselves with the motion.
series eCAADe
last changed 2022/06/07 07:56

For more results click below:

this is page 0show page 1HOMELOGIN (you are user _anon_729214 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002