CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 240

_id cf5c
authors Carpenter, B.
year 1992
title The logic of typed feature structures with applications to unification grammars, logic programs and constraint resolution
source Cambridge Tracts in Theoretical Computer Science, Cambridge University Press
summary This book develops the theory of typed feature structures, a new form of data structure that generalizes both the first-order terms of logic programs and feature-structures of unification-based grammars to include inheritance, typing, inequality, cycles and intensionality. It presents a synthesis of many existing ideas into a uniform framework, which serves as a logical foundation for grammars, logic programming and constraint-based reasoning systems. Throughout the text, a logical perspective is adopted that employs an attribute-value description language along with complete equational axiomatizations of the various systems of feature structures. Efficiency concerns are discussed and complexity and representability results are provided. The application of feature structures to phrase structure grammars is described and completeness results are shown for standard evaluation strategies. Definite clause logic programs are treated as a special case of phrase structure grammars. Constraint systems are introduced and an enumeration technique is given for solving arbitrary attribute-value logic constraints. This book with its innovative approach to data structures will be essential reading for researchers in computational linguistics, logic programming and knowledge representation. Its self-contained presentation makes it flexible enough to serve as both a research tool and a textbook.
series other
last changed 2003/04/23 15:14

_id c434
authors Colajanni, B., Pellitteri, G. and Scianna, A.
year 1992
title Two Approaches to Teaching Computers in Architecture: The Experience in the Faculty of Engineering in Palermo, Italy
doi https://doi.org/10.52842/conf.ecaade.1992.295
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 295-306
summary Teaching the use of computers in architecture poses the same kind of problems as teaching mathematics. To both there are two possible approaches. The first presents the discipline as a tool of which the merely instrumental aspect is emphasized. Teaching is limited to show the results obtainable by existing programs and how to get them. The second approach, on the contrary emphasizes the autonomous nature of the discipline, mathematics as much as computing, on the basis of the convincement that the maximum of instrumental usefulness can be obtained through the knowledge at the highest degree of generality and, then, of abstraction. The first approach changes little in the mind of the student. He simply learns that is possible, and then worthy doing, a certain amount of operations, mainly checks of performances (and not only the control of the aspect, now easy with one of the many existing CAD) or searches of technical informations in some database. The second approach gives the student the consciousness of the manageability of abstract structures of relationships. He acquires then the idea of creating by himself particular structures of relationships and managing them. This can modify the very idea of the design procedure giving the student the consciousness that he can intervene directly in every segment of the design procedure, reshaping it to some extent in a way better suited to the particular problem he is dealing with. Of course this second approach implies learning not only a language but also the capability of coming to terms with languages. And again it is a cultural acquisition that can be very useful when referred to the languages of architecture. Furthermore the capability of simulating on the computer also a small segment of the design process gives the student a better understanding both of the particular problem he is dealing with and of the very nature of design. As for the first effect, it happens whenever a translation is done from a language to another one. One is obliged to get to the core of the matter in order to overcome the difficulties rising from the different bias of the two languages. The second effect comes from the necessity of placing the studied segment in the general flow of the design process. The organisation in a linear sequence of action to be accomplished recursively in an order always varying in any design occasion is an extremely useful exercise to understand the signification and the techniques of formalisation of design problems.
series eCAADe
email
last changed 2022/06/07 07:56

_id 6d1d
authors Daru, R. and Daru, M.
year 1992
title Personal Working Styles in the CMD Studio
doi https://doi.org/10.52842/conf.ecaade.1992.451
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 451-472
summary Normative and problem-solving approaches of architectural design ignore the personality aspects of the designing activity. Every architect approaches projects according to her/his own strategies and tactics. Usually they do not conform to the prescriptive models of design theoreticians. Computer aided design tools should be adapted to their utility within the strategies and tactics of each and every architectural student. We are testing the usefulness of CAAD tools developed by others or ourselves and identifying the needs for missing tools. It is already clear that many CAAD tools reflect the point of view of the programmer about strategies and tactics of designing and that they do not take into account the idiosyncrasies of the end user. Forcing the tools on students breeds the risk of fostering repulsion against ill-adapted tools, and consequently against CMD. Our research group pursues empirical research on working styles of designing by practising architects within the frame of a personality theory of actions. The results indicate that there are three main directions for designing strategies. If we want to take into account the real-world behaviour in design practice within architectural education, this implies the diversification of the exercises we offer to the students in threefold, corresponding with the three directions. To this, we add the didactic options of complementation, compensation and support, depending on what we know about the strong or weak points of the students involved. We have started proposing choices for the exercises of our design morphology studio. Students are offered approaches and tools we consider best adapted to their own working

series eCAADe
email
last changed 2022/06/07 07:55

_id 9f8a
authors Davidow, William H.
year 1992
title The Virtual Corporation: Structuring and Revitalizing the Corporation for the 21St Century
source New York: Harper Collins Publishers
summary The great value of this timely, important book is that it provides an integrated picture of the customer-driven company of the future. We have begun to learn about lean production technology, stripped-down management, worker empowerment, flexible customized manufacturing, and other modern strategies, but Davidow and Malone show for the first time how these ideas are fitting together to create a new kind of corporation and a worldwide business revolution. Their research is fascinating. The authors provide illuminating case studies of American, Japanese, and European companies that have discovered the keys to improved competitiveness, redesigned their businesses and their business relationships, and made extraordinary gains. They also write bluntly and critically about a number of American corporations that are losing market share by clinging to outmoded thinking. Business success in the global marketplace of the future is going to depend upon corporations producing "virtual" products high in added value, rich in variety, and available instantly in response to customer needs. At the heart of this revolution will be fast new information technologies; increased emphasis on quality; accelerated product development; changing management practices, including new alignments between management and labor; and new linkages between company, supplier, and consumer, and between industry and government. The Virtual Corporation is an important cutting-edge book that offers a creative synthesis of the most influential ideas in modern business theory. It has already fired excitement and debate in industry, academia, and government, and it is essential reading for anyone involved in the leadership of America's business and the shaping of America's economic future.
series other
last changed 2003/04/23 15:14

_id 4129
authors Fargas, Josep and Papazian, Pegor
year 1992
title Metaphors in Design: An Experiment with a Frame, Two Lines and Two Rectangles
doi https://doi.org/10.52842/conf.acadia.1992.013
source Mission - Method - Madness [ACADIA Conference Proceedings / ISBN 1-880250-01-2] 1992, pp. 13-22
summary The research we will discuss below originated from an attempt to examine the capacity of designers to evaluate an artifact, and to study the feasibility of replicating a designer's moves intended to make an artifact more expressive of a given quality. We will present the results of an interactive computer experiment, first developed at the MIT Design Research Seminar, which is meant to capture the subject’s actions in a simple design task as a series of successive "moves"'. We will propose that designers use metaphors in their interaction with design artifacts and we will argue that the concept of metaphors can lead to a powerful theory of design activity. Finally, we will show how such a theory can drive the project of building a design system.

When trying to understand how designers work, it is tempting to examine design products in order to come up with the principles or norms behind them. The problem with such an approach is that it may lead to a purely syntactical analysis of design artifacts, failing to capture the knowledge of the designer in an explicit way, and ignoring the interaction between the designer and the evolving design. We will present a theory about design activity based on the observation that knowledge is brought into play during a design task by a process of interpretation of the design document. By treating an evolving design in terms of the meanings and rules proper to a given way of seeing, a designer can reduce the complexity of a task by focusing on certain of its aspects, and can manipulate abstract elements in a meaningful way.

series ACADIA
email
last changed 2022/06/07 07:55

_id ecaade03_473_175_flanagan_neu
id ecaade03_473_175_flanagan_neu
authors Flanagan, Robert H.
year 2003
title Generative Logic in Digital Design
doi https://doi.org/10.52842/conf.ecaade.2003.473
source Digital Design [21th eCAADe Conference Proceedings / ISBN 0-9541183-1-6] Graz (Austria) 17-20 September 2003, pp. 473-484
summary This exploration of early-stage, architectural design pedagogy is in essence, a record of an ongoing transformation underway in architecture, from its practice in the art of geometry of space to its practice in the art of geometry of space-time. A selected series of student experiments, from 1992 to the present, illustrate a progression in architectural theory, from Pythagorean concepts of mathematics and geometry, to the symbolic representation of space and non-linear time in film. The dimensional expansion of space, from xyz to xyz+t (time), represents a tactical and strategic opportunity to incorporate multisensory design variables in architectural practice, as well as in its pedagogy.
keywords Generative; process; derivative; logic; systemic
series eCAADe
email
last changed 2022/06/07 07:51

_id ddss9214
id ddss9214
authors Friedman, A.
year 1993
title A decision-making process for choice of a flexible internal partition option in multi-unit housing using decision theory techniques
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Recent demographic changes have increased the heterogeneity of user groups in the North American housing market. Smaller households (e.g. elderly, single parent) have non-traditional spatial requirements that cannot be accommodated within the conventional house layout. This has created renewed interest in Demountable/Flexible internal partition systems. However, the process by which designers decide which project or user groups are most suited for the use of these systems is quite often complex, non-linear, uncertain and dynamic, since the decisions involve natural processes and human values that are apparently random. The anonymity of users when mass housing projects are conceptualized, and the uncertainty as to the alternative to be selected by the user, given his/her constantly changing needs, are some contributing factors to this effect. Decision Theory techniques, not commonly used by architects, can facilitate the decision-making process through a systematic evaluation of alternatives by means of quantitative methods in order to reduce uncertainty in probabilistic events or in cases when data is insufficient. The author used Decision Theory in the selection of flexible partition systems. The study involved a multi-unit, privately initiated housing project in Montreal, Canada, where real site conditions and costs were used. In this paper, the author outlines the fundamentals of Decision Theory and demonstrates the use of Expected Monetary Value and Weighted Objective Analysis methods and their outcomes in the design of a Montreal housing project. The study showed that Decision Theory can be used as an effective tool in housing design once the designer knows how to collect basic data.
series DDSS
last changed 2003/08/07 16:36

_id 7ce5
authors Gal, Shahaf
year 1992
title Computers and Design Activities: Their Mediating Role in Engineering Education
source Sociomedia, ed. Edward Barret. MIT Press
summary Sociomedia: With all the new words used to describe electronic communication (multimedia, hypertext, cyberspace, etc.), do we need another one? Edward Barrett thinks we do; hence, he coins the term "sociomedia." It is meant to displace a computing economy in which technicity is hypostasized over sociality. Sociomedia, a compilation of twenty-five articles on the theory, design and practice of educational multimedia and hypermedia, attempts to re-value the communicational face of computing. Value, of course, is "ultimately a social construct." As such, it has everything to do with knowledge, power, education and technology. The projects discussed in this book represent the leading edge of electronic knowledge production in academia (not to mention major funding) and are determining the future of educational media. For these reasons, Sociomedia warrants close inspection. Barrett's introduction sets the tone. For him, designing computer media involves hardwiring a mechanism for the social construction of knowledge (1). He links computing to a process of social and communicative interactivity for constructing and desseminating knowledge. Through a mechanistic mapping of the university as hypercontext (a huge network that includes classrooms as well as services and offices), Barrett models intellectual work in such a way as to avoid "limiting definitions of human nature or human development." Education, then, can remain "where it should be--in the human domain (public and private) of sharing ideas and information through the medium of language." By leaving education in a virtual realm (where we can continue to disagree about its meaning and execution), it remains viral, mutating and contaminating in an intellectually healthy way. He concludes that his mechanistic model, by means of its reductionist approach, preserves value (7). This "value" is the social construction of knowledge. While I support the social orientation of Barrett's argument, discussions of value are related to power. I am not referring to the traditional teacher-student power structure that is supposedly dismantled through cooperative and constructivist learning strategies. The power to be reckoned with in the educational arena is foundational, that which (pre)determines value and the circulation of knowledge. "Since each of you reading this paragraph has a different perspective on the meaning of 'education' or 'learning,' and on the processes involved in 'getting an education,' think of the hybris in trying to capture education in a programmable function, in a displayable object, in a 'teaching machine'" (7). Actually, we must think about that hybris because it is, precisely, what informs teaching machines. Moreover, the basic epistemological premises that give rise to such productions are too often assumed. In the case of instructional design, the episteme of cognitive sciences are often taken for granted. It is ironic that many of the "postmodernists" who support electronic hypertextuality seem to have missed Jacques Derrida's and Michel Foucault's "deconstructions" of the epistemology underpinning cognitive sciences (if not of epistemology itself). Perhaps it is the glitz of the technology that blinds some users (qua developers) to the belief systems operating beneath the surface. Barrett is not guilty of reactionary thinking or politics; he is, in fact, quite in line with much American deconstructive and postmodern thinking. The problem arises in that he leaves open the definitions of "education," "learning" and "getting an education." One cannot engage in the production of new knowledge without orienting its design, production and dissemination, and without negotiating with others' orientations, especially where largescale funding is involved. Notions of human nature and development are structural, even infrastructural, whatever the medium of the teaching machine. Although he addresses some dynamics of power, money and politics when he talks about the recession and its effects on the conference, they are readily visible dynamics of power (3-4). Where does the critical factor of value determination, of power, of who gets what and why, get mapped onto a mechanistic model of learning institutions? Perhaps a mapping of contributors' institutions, of the funding sources for the projects showcased and for participation in the conference, and of the disciplines receiving funding for these sorts of projects would help visualize the configurations of power operative in the rising field of educational multimedia. Questions of power and money notwithstanding, Barrett's introduction sets the social and textual thematics for the collection of essays. His stress on interactivity, on communal knowledge production, on the society of texts, and on media producers and users is carried foward through the other essays, two of which I will discuss. Section I of the book, "Perspectives...," highlights the foundations, uses and possible consequences of multimedia and hypertextuality. The second essay in this section, "Is There a Class in This Text?," plays on the robust exchange surrounding Stanley Fish's book, Is There a Text in This Class?, which presents an attack on authority in reading. The author, John Slatin, has introduced electronic hypertextuality and interaction into his courses. His article maps the transformations in "the content and nature of work, and the workplace itself"-- which, in this case, is not industry but an English poetry class (25). Slatin discovered an increase of productive and cooperative learning in his electronically- mediated classroom. For him, creating knowledge in the electronic classroom involves interaction between students, instructors and course materials through the medium of interactive written discourse. These interactions lead to a new and persistent understanding of the course materials and of the participants' relation to the materials and to one another. The work of the course is to build relationships that, in my view, constitute not only the meaning of individual poems, but poetry itself. The class carries out its work in the continual and usually interactive production of text (31). While I applaud his strategies which dismantle traditional hierarchical structures in academia, the evidence does not convince me that the students know enough to ask important questions or to form a self-directing, learning community. Stanley Fish has not relinquished professing, though he, too, espouses the indeterminancy of the sign. By the fourth week of his course, Slatin's input is, by his own reckoning, reduced to 4% (39). In the transcript of the "controversial" Week 6 exchange on Gertrude Stein--the most disliked poet they were discussing at the time (40)--we see the blind leading the blind. One student parodies Stein for three lines and sums up his input with "I like it." Another, finds Stein's poetry "almost completey [sic] lacking in emotion or any artistic merit" (emphasis added). On what grounds has this student become an arbiter of "artistic merit"? Another student, after admitting being "lost" during the Wallace Steven discussion, talks of having more "respect for Stevens' work than Stein's" and adds that Stein's poetry lacks "conceptual significance[, s]omething which people of varied opinion can intelligently discuss without feeling like total dimwits...." This student has progressed from admitted incomprehension of Stevens' work to imposing her (groundless) respect for his work over Stein's. Then, she exposes her real dislike for Stein's poetry: that she (the student) missed the "conceptual significance" and hence cannot, being a person "of varied opinion," intelligently discuss it "without feeling like [a] total dimwit." Slatin's comment is frightening: "...by this point in the semester students have come to feel increasingly free to challenge the instructor" (41). The students that I have cited are neither thinking critically nor are their preconceptions challenged by student-governed interaction. Thanks to the class format, one student feels self-righteous in her ignorance, and empowered to censure. I believe strongly in student empowerment in the classroom, but only once students have accrued enough knowledge to make informed judgments. Admittedly, Slatin's essay presents only partial data (there are six hundred pages of course transcripts!); still, I wonder how much valuable knowledge and metaknowledge was gained by the students. I also question the extent to which authority and professorial dictature were addressed in this course format. The power structures that make it possible for a college to require such a course, and the choice of texts and pedagogy, were not "on the table." The traditional professorial position may have been displaced, but what took its place?--the authority of consensus with its unidentifiable strong arm, and the faceless reign of software design? Despite Slatin's claim that the students learned about the learning process, there is no evidence (in the article) that the students considered where their attitudes came from, how consensus operates in the construction of knowledge, how power is established and what relationship they have to bureaucratic insitutions. How do we, as teaching professionals, negotiate a balance between an enlightened despotism in education and student-created knowledge? Slatin, and other authors in this book, bring this fundamental question to the fore. There is no definitive answer because the factors involved are ultimately social, and hence, always shifting and reconfiguring. Slatin ends his article with the caveat that computerization can bring about greater estrangement between students, faculty and administration through greater regimentation and control. Of course, it can also "distribute authority and power more widely" (50). Power or authority without a specific face, however, is not necessarily good or just. Shahaf Gal's "Computers and Design Activities: Their Mediating Role in Engineering Education" is found in the second half of the volume, and does not allow for a theory/praxis dichotomy. Gal recounts a brief history of engineering education up to the introduction of Growltiger (GT), a computer-assisted learning aid for design. He demonstrates GT's potential to impact the learning of engineering design by tracking its use by four students in a bridge-building contest. What his text demonstrates clearly is that computers are "inscribing and imaging devices" that add another viewpoint to an on-going dialogue between student, teacher, earlier coursework, and other teaching/learning tools. The less proficient students made a serious error by relying too heavily on the technology, or treating it as a "blueprint provider." They "interacted with GT in a way that trusted the data to represent reality. They did not see their interaction with GT as a negotiation between two knowledge systems" (495). Students who were more thoroughly informed in engineering discourses knew to use the technology as one voice among others--they knew enough not simply to accept the input of the computer as authoritative. The less-advanced students learned a valuable lesson from the competition itself: the fact that their designs were not able to hold up under pressure (literally) brought the fact of their insufficient knowledge crashing down on them (and their bridges). They also had, post factum, several other designs to study, especially the winning one. Although competition and comparison are not good pedagogical strategies for everyone (in this case the competitors had volunteered), at some point what we think we know has to be challenged within the society of discourses to which it belongs. Students need critique in order to learn to push their learning into auto-critique. This is what is lacking in Slatin's discussion and in the writings of other avatars of constructivist, collaborative and computer-mediated pedagogies. Obviously there are differences between instrumental types of knowledge acquisition and discoursive knowledge accumulation. Indeed, I do not promote the teaching of reading, thinking and writing as "skills" per se (then again, Gal's teaching of design is quite discursive, if not dialogic). Nevertheless, the "soft" sciences might benefit from "bridge-building" competitions or the re-institution of some forms of agonia. Not everything agonistic is inhuman agony--the joy of confronting or creating a sound argument supported by defensible evidence, for example. Students need to know that soundbites are not sound arguments despite predictions that electronic writing will be aphoristic rather than periodic. Just because writing and learning can be conceived of hypertextually does not mean that rigor goes the way of the dinosaur. Rigor and hypertextuality are not mutually incompatible. Nor is rigorous thinking and hard intellectual work unpleasurable, although American anti-intellectualism, especially in the mass media, would make it so. At a time when the spurious dogmatics of a Rush Limbaugh and Holocaust revisionist historians circulate "aphoristically" in cyberspace, and at a time when knowledge is becoming increasingly textualized, the role of critical thinking in education will ultimately determine the value(s) of socially constructed knowledge. This volume affords the reader an opportunity to reconsider knowledge, power, and new communications technologies with respect to social dynamics and power relationships.
series other
last changed 2003/04/23 15:14

_id c926
authors Laerdal, Arnbjørn O.
year 1992
title Architecture on Cards
doi https://doi.org/10.52842/conf.ecaade.1992.075
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 75-84
summary The ArchiCards project (tentative name) is a prototype of a learning tool in architectural theory and history. It applies this novel technology to give a completely new approach to the acquisition of this kind of information. The aim is to give the user a summary along with an understanding of some of the relations in the world of architecture. Also, it has been an issue to unveil some of the possibilities this technology offers in the teaching of architecture.
series eCAADe
last changed 2022/06/07 07:52

_id ca47
authors Lee, Shu Wan
year 1996
title A Cognitive Approach to Architectural Style Several Characteristics of Design Thinking in Architecture
doi https://doi.org/10.52842/conf.caadria.1996.223
source CAADRIA ‘96 [Proceedings of The First Conference on Computer Aided Architectural Design Research in Asia / ISBN 9627-75-703-9] Hong Kong (Hong Kong) 25-27 April 1996, pp. 223-226
summary Designing is a complicated human behaviour and method, and is often treated as a mysterious "black box” operation in human mind. In the early period as for theory-studying of design thinking, the way of thinking that the researchers took were mostly descriptive discussions. Therefore, they lacked direct and empirical evidence although those studies provided significant exploration of design thinking (Wang, 1995). In recent years as for the study of cognitive science, they have tried to make design "glass box”. That is to try to make the thinking processes embedded in designers publicized. That is also to externalize the design procedure which provided the design studies another theoretical basis of more accurate and deeply researched procedure (Jones, 1992). Hence the studying of design thinking has become more important and the method of designing has also progressed a lot. For example, the classification of the nature of design problem such as ill-defined and well-defined (Newell, Shaw, and Simon, 1967), and different theoretical procedure modes for different disciplines, such as viewing architectural models as conjecture-analysis models and viewing engineering models as analysis-synthesis (Cross, 1991).
series CAADRIA
last changed 2022/06/07 07:52

_id a2e6
authors Liggett, R.S., Mitchell, W.J. and Tan, M.
year 1992
title Multi-Level Analysis and Optimization of Design
source New York: John Wiley & Sons, 1992. pp. 2512-269 : ill. includes bibliography
summary This paper discusses a knowledge-based computer-aided design system, that provides multi-level analysis capabilities, and that automatically propagates constraints on design variables from level to level. It also Supports formulation and solution of optimization problems at different levels, so that a solution can be approached by solving a sequence of appropriately constrained sub-optimization problems. Theory and implementation are discussed, and a detailed case study of application to the design of small house plans is provided
keywords constraints, design, methods, knowledge base, CAD, systems, analysis, optimization, automation, user interface, shape grammars
series CADline
email
last changed 2003/06/02 14:41

_id 181b
authors Liou, Shuenn-Ren
year 1992
title A computer-based framework for analyzing and deriving the morphological structure of architectural designs
source University of Michigan
summary An approach to the acquisition and utilization of knowledge about the morphological structure of notable orthogonal building plans and other two-dimensional compositions is formulated and tested. This approach consists of two levels of abstraction within which the analysis and comparison of existing designs and the derivation of new designs can be undertaken systematically and efficiently. Specifically, the morphological structure of orthogonal building plans and other two-dimensional compositions is conceived as a language defined by shape grammar and architectural grammar corresponding to the geometric and spatial structures of the compositions. Lines constitute the shape grammar and walls and columns the architectural grammar. A computer program named ANADER is designed and implemented using the C++ object-oriented language to describe feasible compositions. It is argued that the gap between morphological analysis and synthesis is bridged partially because the proposed framework facilitates systematic comparisons of the morphological structures of two-dimensional orthogonal compositions and provides insight into the form-making process used to derive them. As an analytical system, the framework contributes to the generation of new and the assessment of existing morphological knowledge. Specifically, it is demonstrated that it is feasible to specify an existing architectural design by a set of universal rule schemata and the sequence of their application. As a generative system, the framework allows many of the tasks involved in the derivation of two-dimensional orthogonal compositions to be carried out. As well, it promotes the use of analytical results. In conclusion, it is argued that the proposed computer-based framework will provide the research and the educator with increasing opportunities for addressing persistent architectural questions in new ways. Of particular interest to this author are questions concerning the decision-making activities involved in form- and space-making as well as the description, classification, and derivation of architecutural form and space. It is suggested that, at least in reference to the cases examined, but probably also in reference to many other morphological classes, these and other related questions can be addressed systematically, efficiently, and fruitfully by using the proposed framework.  
series thesis:PhD
last changed 2003/02/12 22:37

_id ddss9208
id ddss9208
authors Lucardie, G.L.
year 1993
title A functional approach to realizing decision support systems in technical regulation management for design and construction
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Technical building standards defining the quality of buildings, building products, building materials and building processes aim to provide acceptable levels of safety, health, usefulness and energy consumption. However, the logical consistency between these goals and the set of regulations produced to achieve them is often hard to identify. Not only the large quantities of highly complex and frequently changing building regulations to be met, but also the variety of user demands and the steadily increasing technical information on (new) materials, products and buildings have produced a very complex set of knowledge and data that should be taken into account when handling technical building regulations. Integrating knowledge technology and database technology is an important step towards managing the complexity of technical regulations. Generally, two strategies can be followed to integrate knowledge and database technology. The main emphasis of the first strategy is on transferring data structures and processing techniques from one field of research to another. The second approach is concerned exclusively with the semantic structure of what is contained in the data-based or knowledge-based system. The aim of this paper is to show that the second or knowledge-level approach, in particular the theory of functional classifications, is more fundamental and more fruitful. It permits a goal-directed rationalized strategy towards analysis, use and application of regulations. Therefore, it enables the reconstruction of (deep) models of regulations, objects and of users accounting for the flexibility and dynamics that are responsible for the complexity of technical regulations. Finally, at the systems level, the theory supports an effective development of a new class of rational Decision Support Systems (DSS), which should reduce the complexity of technical regulations and restore the logical consistency between the goals of technical regulations and the technical regulations themselves.
series DDSS
last changed 2003/08/07 16:36

_id caadria2020_242
id caadria2020_242
authors Martin Iglesias, Rodrigo, Voto, Cristina and Agra, Rocío
year 2020
title Design in the Age of Dissident Cyborgs - Xenofuturism as caring-curing practices
doi https://doi.org/10.52842/conf.caadria.2020.2.233
source D. Holzer, W. Nakapan, A. Globa, I. Koh (eds.), RE: Anthropocene, Design in the Age of Humans - Proceedings of the 25th CAADRIA Conference - Volume 2, Chulalongkorn University, Bangkok, Thailand, 5-6 August 2020, pp. 233-240
summary This paper synthesizes several years of research in the field of the theory of architecture and design, and its subsequent undergraduate and graduate teaching. Specifically, it is a work that reflects on how architecture and design should face the three most important paradigmatic phenomena of our present and near future. Paradigms as things we think with, rather than as things we think about (Agamben, 2008), or in other words, it matters what ideas we use to think of other ideas (Strathern, 1992). These phenomena refer to environmental, technological and anthropological aspects, and the strategies to cope with them, involving alternate design thinking and practice in which futurabilities and futurizations depart from the displacement generated by post-utopian visions based on dissidence and subalternity.
keywords Chthulucene; Cyborg Design; Dissident Futures; Futurization; Xenofuturism
series CAADRIA
email
last changed 2022/06/07 07:59

_id 1b31
authors Stöckli, Tobi
year 1992
title THE MEASURABLE AND THE UNMEASURABLE OR - FROM FORM TO DESIGN TO EXISTANCE
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part B, pp. 55-62
summary This article discusses the architectural design process from two sides of the spectrum: the formal exercises of experts and the participatory process involving users. The "place" of the full-scale-modelling laboratory at the Federal Institute of Technology in Lausanne is then assessed with respect to this spectrum. It may seem that activities in a full-scale laboratory are closer to the participation process than to formal exercises. However, activities of the full-scale laboratory in Lausanne may best be situated around the middle of the design process. It is clearly within the realm of the measurable (since each construction can easily be measured.) Yet, it does not quite correspond to the real building; it remains an abstraction, a model. And in this quality of abstraction lies the potential to give form to the unmeasurable. It is a tool which allows a transformation of the unmeasurable aspects of an idea into the unmeasurable of existence.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
type normal paper
email
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 15:40

_id 58c5
authors Van Wezel, Ruud
year 1992
title MOCK-UP SYSTEM WAGENINGEN: DEVELOPMENT, LIMITATION AND FUTURE
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part A, pp. 15-18
summary A brief description of the development of the Mock-up System (MUS) in the context of the Wageningen training program. The students are first taught some keywords in understanding of the building process. They are then trained to express how they want to live (theory) and later on they confront themselves with what they have built in the MUS (practice) . Besides being an educational tool, the MUS is used for pre-building evaluation and research questions. The drawbacks or limitations of the system (outdoor reality versus indoor simulation) and future use by different target groups are also discussed in this paper. The power of the MUS is, and will continue to be, the concrete building of communicational results and the generation of communication by doing so.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 15:30

_id f1f2
authors Watt, Alan
year 1992
title Advanced animation and rendering techniques
source New York: ACM Press
summary This book is an exposition of state-of-the-art techniques in rendering and animation. It provides a unique synthesis of techniques and theory. Four sections describe: Basics, Theoretical Foundations, Advanced Rendering Techniques, and Advanced Animation Techniques. Each technique is illustrated with a series of full color frames showing the development of the example. Many code examples and some complete implementations are given in C for interesting and advanced algorithms such as soft shadows and marching cubes.
series other
last changed 2003/04/23 15:14

_id 3ff5
authors Abbo, I.A., La Scalea, L., Otero, E. and Castaneda, L.
year 1992
title Full-Scale Simulations as Tool for Developing Spatial Design Ability
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part C, pp. 7-10
summary Spatial Design Ability has been defined as the capability to anticipate effects (psychological impressions on potential observers or users) produced by mental manipulation of elements of architectural or urban spaces. This ability, of great importance in choosing the appropriate option during the design process, is not specifically developed in schools of architecture and is partially obtained as a by-product of drawing, designing or architectural criticism. We use our Laboratory as a tool to present spaces to people so that they can evaluate them. By means of a series of exercises, students confront their anticipations with the psychological impressions produced in other people. For this occasion, we present an experience in which students had to propose a space for an exhibition hag in which architectural projects (student thesis) were to be shown. Following the Spatial Design Ability Development Model which we have been using for several years, students first get acquainted with the use of evaluation instruments for psychological impressions as well as with research methodology. In this case, due to the short period available, we reduced research to investigate the effects produced by the manipulation of only 2 independents variables: students manipulated first the form of the roof, walls and interiors elements, secondly, color and texture of those elements. They evaluated spatial quality, character and the other psychological impressions that manipulations produced in people. They used three dimensional scale models 1/10 and 1/1.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
email
more http://info.tuwien.ac.at/efa
last changed 2003/08/25 10:12

_id 6208
authors Abou-Jaoude, Georges
year 1992
title To Master a Tool
source Proceedings of the 4rd European Full-Scale Modelling Conference / Lausanne (Switzerland) 9-12 September 1992, Part B, p. 15
summary The tool here is the computer or to be precise, a unit that includes the computer, the peripherals and the software needed to fulfill a task. These tools are getting very sophisticated and user interfaces extremly friendly, therefore it is very easy to become the slave of such electronic tools and reach self satisfaction with strait forward results and attractive images. In order to master and not to become slaves of sophisticated tools, a very solid knowledge of related fields or domains of application becomes necessary. In the case of this seminar, full scale modelling, is a way to understand the relation between a mental model and it's full-scale modelling, it is a way of communicating what is in a designers mind. Computers and design programs can have the same goal, rather than chosing one method or the other let us try to say how important it is today to complement designing with computer with other means and media such as full scale modelling, and what computer modelling and simulation can bring to full scale modelling or other means.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
more http://info.tuwien.ac.at/efa
last changed 2003/08/25 10:12

_id acadia06_455
id acadia06_455
authors Ambach, Barbara
year 2006
title Eve’s Four Faces interactive surface configurations
doi https://doi.org/10.52842/conf.acadia.2006.455
source Synthetic Landscapes [Proceedings of the 25th Annual Conference of the Association for Computer-Aided Design in Architecture] pp. 455-460
summary Eve’s Four Faces consists of a series of digitally animated and interactive surfaces. Their content and structure are derived from a collection of sources outside the conventional boundaries of architectural research, namely psychology and the broader spectrum of arts and culture.The investigation stems from a psychological study documenting the attributes and social relationships of four distinct personality prototypes: the Individuated, the Traditional, the Conflicted, and the Assured (York and John 1992). For the purposes of this investigation, all four prototypes are assumed to be inherent, to certain degrees, in each individual. However, the propensity towards one of the prototypes forms the basis for each individual’s “personality structure.” The attributes, social implications and prospects for habitation have been translated into animations and surfaces operating within A House for Eve’s Four Faces. The presentation illustrates the potential for constructed surfaces to be configured and transformed interactively, responding to the needs and qualities associated with each prototype. The intention is to study the effects of each configuration and how each configuration may be therapeutic in supporting, challenging or altering one’s personality as it oscillates and shifts through the four prototypical conditions.
series ACADIA
email
last changed 2022/06/07 07:54

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 11HOMELOGIN (you are user _anon_604777 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002