CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 245

_id cb5a
authors Oxman, Rivka E.
year 1992
title Multiple Operative and Interactive Modes in Knowledge-Based Design Systems
source New York: John Wiley & Sons, 1992. pp. 125-143 : ill. includes bibliography
summary A conceptual basis for the development of an expert system which is capable of integrating various modes of generation and evaluation in design is presented. This approach is based upon two sets of reasoning processes in the design system. The first enables a mapping between design requirements and solution descriptions in a generative mode of design; and the second enables a mapping between solution descriptions and performance evaluation in an evaluative and predictive mode. This concept supports a formal framework necessary for a knowledge-based design system to operate in a design partnership relation with the designer. Another fundamental concept in expert systems for design, dual direction interpretation between graphic and textual modes, is presented and elaborated. This encoding of knowledge behind the geometrical representation can be achieved in knowledge- based design systems by the development of a 'semantic interpreter' which supports a dual direction mapping process employing a geometrical knowledge, typological knowledge and evaluative knowledge. An implemented expert system for design, PREDIKT, demonstrates these concepts in the domain of kitchen design. It provides the user with a choice of alternative modes of interaction, such as: a 'design critic' for the evaluation of a design, a 'design generator' for the generation of a design, or a 'design critic-generator' for the completion of partial solutions
keywords architecture, knowledge base, design, systems, expert systems
series CADline
email
last changed 2003/06/02 10:24

_id 7ce5
authors Gal, Shahaf
year 1992
title Computers and Design Activities: Their Mediating Role in Engineering Education
source Sociomedia, ed. Edward Barret. MIT Press
summary Sociomedia: With all the new words used to describe electronic communication (multimedia, hypertext, cyberspace, etc.), do we need another one? Edward Barrett thinks we do; hence, he coins the term "sociomedia." It is meant to displace a computing economy in which technicity is hypostasized over sociality. Sociomedia, a compilation of twenty-five articles on the theory, design and practice of educational multimedia and hypermedia, attempts to re-value the communicational face of computing. Value, of course, is "ultimately a social construct." As such, it has everything to do with knowledge, power, education and technology. The projects discussed in this book represent the leading edge of electronic knowledge production in academia (not to mention major funding) and are determining the future of educational media. For these reasons, Sociomedia warrants close inspection. Barrett's introduction sets the tone. For him, designing computer media involves hardwiring a mechanism for the social construction of knowledge (1). He links computing to a process of social and communicative interactivity for constructing and desseminating knowledge. Through a mechanistic mapping of the university as hypercontext (a huge network that includes classrooms as well as services and offices), Barrett models intellectual work in such a way as to avoid "limiting definitions of human nature or human development." Education, then, can remain "where it should be--in the human domain (public and private) of sharing ideas and information through the medium of language." By leaving education in a virtual realm (where we can continue to disagree about its meaning and execution), it remains viral, mutating and contaminating in an intellectually healthy way. He concludes that his mechanistic model, by means of its reductionist approach, preserves value (7). This "value" is the social construction of knowledge. While I support the social orientation of Barrett's argument, discussions of value are related to power. I am not referring to the traditional teacher-student power structure that is supposedly dismantled through cooperative and constructivist learning strategies. The power to be reckoned with in the educational arena is foundational, that which (pre)determines value and the circulation of knowledge. "Since each of you reading this paragraph has a different perspective on the meaning of 'education' or 'learning,' and on the processes involved in 'getting an education,' think of the hybris in trying to capture education in a programmable function, in a displayable object, in a 'teaching machine'" (7). Actually, we must think about that hybris because it is, precisely, what informs teaching machines. Moreover, the basic epistemological premises that give rise to such productions are too often assumed. In the case of instructional design, the episteme of cognitive sciences are often taken for granted. It is ironic that many of the "postmodernists" who support electronic hypertextuality seem to have missed Jacques Derrida's and Michel Foucault's "deconstructions" of the epistemology underpinning cognitive sciences (if not of epistemology itself). Perhaps it is the glitz of the technology that blinds some users (qua developers) to the belief systems operating beneath the surface. Barrett is not guilty of reactionary thinking or politics; he is, in fact, quite in line with much American deconstructive and postmodern thinking. The problem arises in that he leaves open the definitions of "education," "learning" and "getting an education." One cannot engage in the production of new knowledge without orienting its design, production and dissemination, and without negotiating with others' orientations, especially where largescale funding is involved. Notions of human nature and development are structural, even infrastructural, whatever the medium of the teaching machine. Although he addresses some dynamics of power, money and politics when he talks about the recession and its effects on the conference, they are readily visible dynamics of power (3-4). Where does the critical factor of value determination, of power, of who gets what and why, get mapped onto a mechanistic model of learning institutions? Perhaps a mapping of contributors' institutions, of the funding sources for the projects showcased and for participation in the conference, and of the disciplines receiving funding for these sorts of projects would help visualize the configurations of power operative in the rising field of educational multimedia. Questions of power and money notwithstanding, Barrett's introduction sets the social and textual thematics for the collection of essays. His stress on interactivity, on communal knowledge production, on the society of texts, and on media producers and users is carried foward through the other essays, two of which I will discuss. Section I of the book, "Perspectives...," highlights the foundations, uses and possible consequences of multimedia and hypertextuality. The second essay in this section, "Is There a Class in This Text?," plays on the robust exchange surrounding Stanley Fish's book, Is There a Text in This Class?, which presents an attack on authority in reading. The author, John Slatin, has introduced electronic hypertextuality and interaction into his courses. His article maps the transformations in "the content and nature of work, and the workplace itself"-- which, in this case, is not industry but an English poetry class (25). Slatin discovered an increase of productive and cooperative learning in his electronically- mediated classroom. For him, creating knowledge in the electronic classroom involves interaction between students, instructors and course materials through the medium of interactive written discourse. These interactions lead to a new and persistent understanding of the course materials and of the participants' relation to the materials and to one another. The work of the course is to build relationships that, in my view, constitute not only the meaning of individual poems, but poetry itself. The class carries out its work in the continual and usually interactive production of text (31). While I applaud his strategies which dismantle traditional hierarchical structures in academia, the evidence does not convince me that the students know enough to ask important questions or to form a self-directing, learning community. Stanley Fish has not relinquished professing, though he, too, espouses the indeterminancy of the sign. By the fourth week of his course, Slatin's input is, by his own reckoning, reduced to 4% (39). In the transcript of the "controversial" Week 6 exchange on Gertrude Stein--the most disliked poet they were discussing at the time (40)--we see the blind leading the blind. One student parodies Stein for three lines and sums up his input with "I like it." Another, finds Stein's poetry "almost completey [sic] lacking in emotion or any artistic merit" (emphasis added). On what grounds has this student become an arbiter of "artistic merit"? Another student, after admitting being "lost" during the Wallace Steven discussion, talks of having more "respect for Stevens' work than Stein's" and adds that Stein's poetry lacks "conceptual significance[, s]omething which people of varied opinion can intelligently discuss without feeling like total dimwits...." This student has progressed from admitted incomprehension of Stevens' work to imposing her (groundless) respect for his work over Stein's. Then, she exposes her real dislike for Stein's poetry: that she (the student) missed the "conceptual significance" and hence cannot, being a person "of varied opinion," intelligently discuss it "without feeling like [a] total dimwit." Slatin's comment is frightening: "...by this point in the semester students have come to feel increasingly free to challenge the instructor" (41). The students that I have cited are neither thinking critically nor are their preconceptions challenged by student-governed interaction. Thanks to the class format, one student feels self-righteous in her ignorance, and empowered to censure. I believe strongly in student empowerment in the classroom, but only once students have accrued enough knowledge to make informed judgments. Admittedly, Slatin's essay presents only partial data (there are six hundred pages of course transcripts!); still, I wonder how much valuable knowledge and metaknowledge was gained by the students. I also question the extent to which authority and professorial dictature were addressed in this course format. The power structures that make it possible for a college to require such a course, and the choice of texts and pedagogy, were not "on the table." The traditional professorial position may have been displaced, but what took its place?--the authority of consensus with its unidentifiable strong arm, and the faceless reign of software design? Despite Slatin's claim that the students learned about the learning process, there is no evidence (in the article) that the students considered where their attitudes came from, how consensus operates in the construction of knowledge, how power is established and what relationship they have to bureaucratic insitutions. How do we, as teaching professionals, negotiate a balance between an enlightened despotism in education and student-created knowledge? Slatin, and other authors in this book, bring this fundamental question to the fore. There is no definitive answer because the factors involved are ultimately social, and hence, always shifting and reconfiguring. Slatin ends his article with the caveat that computerization can bring about greater estrangement between students, faculty and administration through greater regimentation and control. Of course, it can also "distribute authority and power more widely" (50). Power or authority without a specific face, however, is not necessarily good or just. Shahaf Gal's "Computers and Design Activities: Their Mediating Role in Engineering Education" is found in the second half of the volume, and does not allow for a theory/praxis dichotomy. Gal recounts a brief history of engineering education up to the introduction of Growltiger (GT), a computer-assisted learning aid for design. He demonstrates GT's potential to impact the learning of engineering design by tracking its use by four students in a bridge-building contest. What his text demonstrates clearly is that computers are "inscribing and imaging devices" that add another viewpoint to an on-going dialogue between student, teacher, earlier coursework, and other teaching/learning tools. The less proficient students made a serious error by relying too heavily on the technology, or treating it as a "blueprint provider." They "interacted with GT in a way that trusted the data to represent reality. They did not see their interaction with GT as a negotiation between two knowledge systems" (495). Students who were more thoroughly informed in engineering discourses knew to use the technology as one voice among others--they knew enough not simply to accept the input of the computer as authoritative. The less-advanced students learned a valuable lesson from the competition itself: the fact that their designs were not able to hold up under pressure (literally) brought the fact of their insufficient knowledge crashing down on them (and their bridges). They also had, post factum, several other designs to study, especially the winning one. Although competition and comparison are not good pedagogical strategies for everyone (in this case the competitors had volunteered), at some point what we think we know has to be challenged within the society of discourses to which it belongs. Students need critique in order to learn to push their learning into auto-critique. This is what is lacking in Slatin's discussion and in the writings of other avatars of constructivist, collaborative and computer-mediated pedagogies. Obviously there are differences between instrumental types of knowledge acquisition and discoursive knowledge accumulation. Indeed, I do not promote the teaching of reading, thinking and writing as "skills" per se (then again, Gal's teaching of design is quite discursive, if not dialogic). Nevertheless, the "soft" sciences might benefit from "bridge-building" competitions or the re-institution of some forms of agonia. Not everything agonistic is inhuman agony--the joy of confronting or creating a sound argument supported by defensible evidence, for example. Students need to know that soundbites are not sound arguments despite predictions that electronic writing will be aphoristic rather than periodic. Just because writing and learning can be conceived of hypertextually does not mean that rigor goes the way of the dinosaur. Rigor and hypertextuality are not mutually incompatible. Nor is rigorous thinking and hard intellectual work unpleasurable, although American anti-intellectualism, especially in the mass media, would make it so. At a time when the spurious dogmatics of a Rush Limbaugh and Holocaust revisionist historians circulate "aphoristically" in cyberspace, and at a time when knowledge is becoming increasingly textualized, the role of critical thinking in education will ultimately determine the value(s) of socially constructed knowledge. This volume affords the reader an opportunity to reconsider knowledge, power, and new communications technologies with respect to social dynamics and power relationships.
series other
last changed 2003/04/23 15:14

_id e7c8
authors Kalisperis, Loukas N., Steinman, Mitch and Summers, Luis H.
year 1992
title Design Knowledge, Environmental Complexity in Nonorthogonal Space
source New York: John Wiley & Sons, 1992. pp. 273-291 : ill. includes bibliography
summary Mechanization and industrialization of society has resulted in most people spending the greater part of their lives in enclosed environments. Optimal design of indoor artificial climates is therefore of increasing importance. Wherever artificial climates are created for human occupation, the aim is that the environment be designed so that individuals are in thermal comfort. Current design methodologies for radiant panel heating systems do not adequately account for the complexities of human thermal comfort, because they monitor air temperature alone and do not account for thermal neutrality in complex enclosures. Thermal comfort for a person is defined as that condition of mind which expresses satisfaction with the thermal environment. Thermal comfort is dependent on Mean Radiant Temperature and Operative Temperature among other factors. In designing artificial climates for human occupancy the interaction of the human with the heated surfaces as well the surface-to-surface heat exchange must be accounted for. Early work in the area provided an elaborate and difficult method for calculating radiant heat exchange for simplistic and orthogonal enclosures. A new improved method developed by the authors for designing radiant panel heating systems based on human thermal comfort and mean radiant temperature is presented. Through automation and elaboration this method overcomes the limitations of the early work. The design procedure accounts for human thermal comfort in nonorthogonal as well as orthogonal spaces based on mean radiant temperature prediction. The limitation of simplistic orthogonal geometries has been overcome with the introduction of the MRT-Correction method and inclined surface-to-person shape factor methodology. The new design method increases the accuracy of calculation and prediction of human thermal comfort and will allow designers to simulate complex enclosures utilizing the latest design knowledge of radiant heat exchange to increase human thermal comfort
keywords applications, architecture, building, energy, systems, design, knowledge
series CADline
last changed 2003/06/02 10:24

_id a081
authors Greenberg S., Roseman M. and Webster, D.
year 1992
title Issues and Experiences Designing and Implementing Two Group Drawing Tools
source Readings in Groupware, 609-620
summary Groupware designers are now developing multi-user equivalents of popular paint and draw applications. Their job is not an easy one. First, human factors issues peculiar to group interaction appear that, if ignored, seriously limit the usability of the group tool. Second, implementation is fraught with considerable hurdles. This paper describes the issues and experiences we have met and handled in the design of two systems supporting remote real time group interaction: GroupSketch, a multi-user sketchpad; and GroupDraw, an object-based multi-user draw package. On the human factors side, we summarize empirically-derived design principles that we believe are critical to building useful and usable collaborative drawing tools. On the implementation side, we describe our experiences with replicated versus centralized architectures, schemes for participant registration, multiple cursors, network requirements, and the structure of the drawing primitives.
series other
last changed 2003/04/23 15:50

_id b2f9
id b2f9
authors Bhzad Sidawi and Neveen Hamza
year 2012
title INTELLIGENT KNOWLEDGE-BASED REPOSITORY TO SUPPORT INFORMED DESIGN DECISION MAKING
source ITCON journal
summary Research highlights that architectural design is a social phenomenon that is underpinned by critical analysis of design precedents and the social interaction between designers including negotiation, collaboration and communication. CAAD systems are continuously developing as essential design tools in formulating and developing ideas. Researchers such as (Rosenman, Gero and Oxman 1992) have suggested suggest that knowledge based systems can be integrated with CAAD systems to provide design knowledge that would enable recalling design precedents that maybe linked to the design constraints. Currently CAAD systems are user centric being focused on architects rather than the end product. The systems provide limited assistance in the production of innovative design. Furthermore, the attention of the designers of knowledge based systems is providing a repository rather than a system that is capable to initiate innovation. Most of the CAAD systems have web communication tools that enable designers to communicate their design ideas with colleagues and partners in business. However, none of these systems have the capability to capture useful knowledge from the design negotiations. Students of the third to fifth year at College of Architecture, University of Dammam were surveyed and interviewed to find out how far design tools, communications and resources would impact the production of innovative design projects. The survey results show that knowledge extracted from design negotiations would impact the innovative design outcome. It highlights also that present design precedents are not very helpful and design negotiations between students, tutors and other students are not documented thus fully incorporated into the design scheme. The paper argues that the future CAAD systems should be capable to recognize innovative design precedents, and incorporate knowledge that is resulted from design negotiations. This would help students to gain a critical mass of knowledge that would underpin informed design decisions.
series journal paper
type normal paper
email
more http://www.itcon.org/cgi-bin/works/Show?2012_20
last changed 2012/09/19 13:41

_id ddss9209
id ddss9209
authors De Gelder, J.T. and Lucardie, G.L.
year 1993
title Knowledge and data modelling in cad/cam applications
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Modelling knowledge and data in CAD/CAM applications is complex because different goals and contexts have to be taken into account. This complexity makes particular demands upon representation formalisms. Today many modelling tools are based on record structures. By analyzing the requirements for a product model of a portal structure in steel, this paper shows that in many situations record structures are not well suited as a representation formalism for storing knowledge and data in CAD/CAM applications. This is illustrated by performing a knowledge-level analysis of the knowledge and data generated in the design and manufacturing process of a portal structure in steel.
series DDSS
last changed 2003/08/07 16:36

_id 4129
authors Fargas, Josep and Papazian, Pegor
year 1992
title Metaphors in Design: An Experiment with a Frame, Two Lines and Two Rectangles
doi https://doi.org/10.52842/conf.acadia.1992.013
source Mission - Method - Madness [ACADIA Conference Proceedings / ISBN 1-880250-01-2] 1992, pp. 13-22
summary The research we will discuss below originated from an attempt to examine the capacity of designers to evaluate an artifact, and to study the feasibility of replicating a designer's moves intended to make an artifact more expressive of a given quality. We will present the results of an interactive computer experiment, first developed at the MIT Design Research Seminar, which is meant to capture the subject’s actions in a simple design task as a series of successive "moves"'. We will propose that designers use metaphors in their interaction with design artifacts and we will argue that the concept of metaphors can lead to a powerful theory of design activity. Finally, we will show how such a theory can drive the project of building a design system.

When trying to understand how designers work, it is tempting to examine design products in order to come up with the principles or norms behind them. The problem with such an approach is that it may lead to a purely syntactical analysis of design artifacts, failing to capture the knowledge of the designer in an explicit way, and ignoring the interaction between the designer and the evolving design. We will present a theory about design activity based on the observation that knowledge is brought into play during a design task by a process of interpretation of the design document. By treating an evolving design in terms of the meanings and rules proper to a given way of seeing, a designer can reduce the complexity of a task by focusing on certain of its aspects, and can manipulate abstract elements in a meaningful way.

series ACADIA
email
last changed 2022/06/07 07:55

_id 83f7
authors Fenves, Stephen J., Flemming, Ulrich and Hendrickson, Craig (et al)
year 1992
title Performance Evaluation in an Integrated Software Environment for Building Design and Construction Planning
source New York: John Wiley & Sons, 1992. pp. 159-169 : ill. includes bibliography
summary In this paper the authors describe the role of performance evaluation in the Integrated Software Environment for Building Design and Construction Planning (IBDE), which is a testbed for examining integration issues in the same domain. Various processes in IBDE deal with the spatial configuration, structural design, and construction planning of high-rise office buildings. Performance evaluations occur within these processes based on different representation schemes and control mechanisms for the handling of performance knowledge. Within this multiprocess environment, opportunities also exist for performance evaluation across disciplines through design critics
keywords evaluation, performance, integration, systems, building, design, construction, architecture, planning, structures, representation, control
series CADline
email
last changed 2003/06/02 10:24

_id cc2f
authors Jog, Bharati
year 1992
title Evaluation of Designs for Energy Performance Using A Knowledge-Based System
source New York: John Wiley & Sons, 1992. pp. 293-304 : ill. includes a bibliography
summary Principles of knowledge-based (or expert) systems have been applied in different knowledge-rich domains such as geology, medicine, and very large scale integrated circuits (VLSI). There have been some efforts to develop expert systems for evaluation and prediction of architectural designs in this decade. This paper presents a prototype system, Energy Expert, which quickly computes the approximate yearly energy performance of a building design, analyzes the energy performance, and gives advice on possible ways of improving the design. These modifications are intended to make the building more energy efficient and help cut down on heating and cooling costs. The system is designed for the schematic design phase of an architectural project. Also discussed briefly is the reasoning behind developing such a system for the schematic design rather than the final design phase
keywords expert systems, energy, evaluation, performance, knowledge base, architecture, reasoning, programming, prediction
series CADline
last changed 1999/02/12 15:08

_id ddss9208
id ddss9208
authors Lucardie, G.L.
year 1993
title A functional approach to realizing decision support systems in technical regulation management for design and construction
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary Technical building standards defining the quality of buildings, building products, building materials and building processes aim to provide acceptable levels of safety, health, usefulness and energy consumption. However, the logical consistency between these goals and the set of regulations produced to achieve them is often hard to identify. Not only the large quantities of highly complex and frequently changing building regulations to be met, but also the variety of user demands and the steadily increasing technical information on (new) materials, products and buildings have produced a very complex set of knowledge and data that should be taken into account when handling technical building regulations. Integrating knowledge technology and database technology is an important step towards managing the complexity of technical regulations. Generally, two strategies can be followed to integrate knowledge and database technology. The main emphasis of the first strategy is on transferring data structures and processing techniques from one field of research to another. The second approach is concerned exclusively with the semantic structure of what is contained in the data-based or knowledge-based system. The aim of this paper is to show that the second or knowledge-level approach, in particular the theory of functional classifications, is more fundamental and more fruitful. It permits a goal-directed rationalized strategy towards analysis, use and application of regulations. Therefore, it enables the reconstruction of (deep) models of regulations, objects and of users accounting for the flexibility and dynamics that are responsible for the complexity of technical regulations. Finally, at the systems level, the theory supports an effective development of a new class of rational Decision Support Systems (DSS), which should reduce the complexity of technical regulations and restore the logical consistency between the goals of technical regulations and the technical regulations themselves.
series DDSS
last changed 2003/08/07 16:36

_id 975e
authors Pearce, M. and Goel, A. (et al.)
year 1992
title Case-Based Design support: A case study in architectural design
source IEEE Expert 7(5): 14-20
summary Archie, a small computer-based library of architectural design cases, is described. Archie helps architects in the high-level task of conceptual design as opposed to low-level tasks such as drawing and drafting, numerical calculations, and constraint propagation. Archie goes beyond supporting architects in design proposal and critiquing. It acts as a shared external memory that supports two kinds of design collaboration. First, by including enough knowledge about the goals, plans, outcomes, and lessons of past cases, it lets the designer access the work of previous architects. Second, by providing access to the perspectives of domain experts via the domain models, Archie helps architects anticipate and accommodate experts' views on evolving designs. The lessons learned about building large case-based systems to support real-world decision making in developing Archie are discussed.
series journal paper
last changed 2003/04/23 15:14

_id daff
authors Richens, P.
year 1994
title CAD Research at the Martin Centre
source Automation in Construction, No. 3
summary The Martin Centre CADLAB has recently been established to investigate software techniques that could be of practical importance to architects within the next five years. In common with most CAD researchers, we are interested in the earlier, conceptual, stages of design, where commercial CAD systems have had little impact. Our approach is not Knowledge-Based, but rather focuses on using the computer as a medium for design and communication. This leads to a concentration on apparently superficial aspects such as visual appearance, the dynamics of interaction, immediate feedback, plasticity. We try to avoid building-in theoretical attitudes, and to reduce the semantic content of our systems to a low level on the basis that flexibility and intelligence are inversely related; and that flexibility is more important. The CADLAB became operational in January 1992. First year work in three areas – building models, experiencing architecture, and making drawings – is discussed.
series journal
email
more http://www.arct.cam.ac.uk/research/pubs/pdfs/rich94a.pdf
last changed 2000/03/05 19:05

_id c804
authors Richens, P.
year 1994
title Does Knowledge really Help?
source G. Carrara and Y.E. Kalay (Eds.), Knowledge-Based Computer-Aided Architectural Design, Elsevier
summary The Martin Centre CADLAB has recently been established to investigate software techniques that could be of practical importance to architects within the next five years. In common with most CAD researchers, we are interested in the earlier, conceptual, stages of design, where commercial CAD systems have had little impact. Our approach is not Knowledge-Based, but rather focuses on using the computer as a medium for design and communication. This leads to a concentration on apparently superficial aspects such as visual appearance, the dynamics of interaction, immediate feedback, plasticity. We try to avoid building-in theoretical attitudes, and to reduce the semantic content of our systems to a low level on the basis that flexibility and intelligence are inversely related; and that flexibility is more important. The CADLAB became operational in January 1992. First year work in three areas – building models, experiencing architecture, and making drawings – is discussed.
series other
more http://www.arct.cam.ac.uk/research/pubs/
last changed 2003/03/05 13:19

_id fd02
authors Tsou, Jin-Yeu
year 1992
title Using conceptual modelling and an object-oriented environment to support building cost control during early design
source College of Architecture and Urban Planning, University of Michigan
summary This research investigated formal information modelling techniques and the object-oriented knowledge representation on the domain of building cost control during early design stages. The findings contribute to an understanding of the advantages and disadvantages of applying formal modelling techniques to the analysis of architectural problems and the representation of domain knowledge in an object-oriented environment. In this study, information modelling techniques were reviewed, formal information analysis was performed, a conceptual model based on the cost control problem domain was created, a computational model based on the object-oriented approach was developed, a mechanism to support information broadcasting for representing interrelationships was implemented, and an object-oriented cost analysis system for early design (OBCIS) was demonstrated. The conceptual model, based on the elemental proposition analysis of NIAM, supports a formal approach for analyzing the problem domain; the analysis results are represented by high-level graphical notations, based on the AEC Building System Model, to visually display the information framework of the domain. The conceptual model provides an intermediate step between the system designer's view of the domain and the internal representation of the implementation platform. The object-oriented representation provides extensive data modelling abilities to help system designers intuitively represent the semantics of the problem domain. The object-oriented representation also supports more structured and integrated modules than conventional programming approaches. Although there are many advantages to applying this technique to represent the semantics of cost control knowledge, there are several issues which need to be considered: no single satisfactory classification method can be directly applied; object-oriented systems are difficult to learn; and designing reusable classes is difficult. The dependency graph and information broadcasting implemented in this research is an attempt to represent the interrelationships between domain objects. The mechanism allows users to explicitly define the interrelationships, based on semantic requirements, among domain objects. In the conventional approach, these relationships are directly interpreted by system designers and intertwined into the programming code. There are several issues which need to be studied further: indirect dependency relationship, conflict resolution, and request-update looping based on least-commitment approach.
series thesis:PhD
email
last changed 2003/02/12 22:37

_id ddss9201
id ddss9201
authors Van Bakel, A.P.M.
year 1993
title Personality assessment in regard to design strategies
source Timmermans, Harry (Ed.), Design and Decision Support Systems in Architecture (Proceedings of a conference held in Mierlo, the Netherlands in July 1992), ISBN 0-7923-2444-7
summary This paper discusses some preliminary results of several knowledge-acquisition and documentation-structuring techniques that were used to assess the working styles of architects. The focus of this assessment was on their strategic design behaviour. Hettema's Interactive Personality Model (Hettema 1979, 1989) was used to explain and interpret these results. The methods used to acquire the necessary data are protocol analysis, card sorting and interviews. The results suggest that at least three parameters can be used to explain and differentiate the strategic design behaviour of architects. These parameters are S (site-oriented), B (brief-oriented) and C (concept-oriented). A priority hierarchy of these parameters reveals six major distinguishable working styles. These results are captured in a new design model that can be used in data bank implementations.
series DDSS
last changed 2003/08/07 16:36

_id cbed
authors Yakubu, G.S.
year 1994
title Maximising the Benefits of CAD Systems in Architectural Education
doi https://doi.org/10.52842/conf.ecaade.1994.x.u8n
source The Virtual Studio [Proceedings of the 12th European Conference on Education in Computer Aided Architectural Design / ISBN 0-9523687-0-6] Glasgow (Scotland) 7-10 September 1994, p. 228
summary The positive impact of Computer Aided Design (CAD) in professional architectural practice has been in focus in recent times but relatively little has been written on its significance in the education of the contemporary architect. It is common knowledge that the profession of architecture is currently undergoing enormous strains as it battles to keep abreast of trends and developments in a period of series of rapid advancement in science, technology and management (RIBA, 1992). Whilst attempts are being made to redress the shortcomings of the profession in the above context, the requirements for architectural education are yet to forge a coherent strategy for the implementation of CAD/IT in the curriculum of schools of architecture. In almost every other field, including engineering, medicine and the humanities, computing application to problem-solving and decision-making is seen as a way forward as we move into 21st century. Architectural education must integrate CAD/IT into the teaching of core modules that give the architect distinctive competence: studio design. That is one of the best ways of doing justice to the education of the architect of today and the future. Some approaches to the teaching of CAD in schools of architecture have been touched upon in the recent past. Building upon this background as well as an understanding of the nature of design teaching/learning, this paper examines ways of maximising the benefits of CAD systems in architectural education and of bringing computer aided designing into the studio not only to enhance design thinking and creativity but also to support interactive processes. In order to maximise or optimise any function, one approach is to use the hard systems methodology which utilises analytic, analogic and iconic models to show the effect of those factors which are significant for the purposes being considered. The other approach is to use the soft systems methodology in which the analysis encompasses the concept of a human activity system as a means of improving a situation. The use of soft systems methodology is considered more appropriate for dealing with the problem of design which is characterised by a flux of interacting events and ideas that unroll through time. The paper concludes that the main impediment to maximising the benefits of CAD systems in architectural education is not only the inappropriate definition of the objectives for the implementation of CAD education but also that the control subsystems are usually ill-structured and relatively poorly defined. Schools must attempt to define a coherent and consistent policy on the use of CAD systems as an integral part of studio design and evolve an in-house strategic and operational controls that enable the set objectives to be met. Furthermore, it is necessary to support the high level of productivity from CAD systems with a more efficient management system, especially in dealing with communication, data sharing via relational database, co-ordination and integration. Finally, the use of soft systems methodology is recommended as the way forward to optimising CAD systems in design education as it would provide continuous improvements while maintaining their productive value.

series eCAADe
last changed 2022/06/07 07:50

_id 0ff4
authors Woodbury, R.F., Radford, A.D., Taplin, P.N. and Coppins, S.A.
year 1992
title Tartan Worlds: A Generative Symbol Grammar System
doi https://doi.org/10.52842/conf.acadia.1992.211
source Mission - Method - Madness [ACADIA Conference Proceedings / ISBN 1-880250-01-2] 1992, pp. 211-220
summary Tartan Worlds is a highly interactive, generative symbol grammar system that generates designs as two dimensional configurations of symbols on tartan grids. It operates with multiple graphical rule sets on multiple design worlds. Designers can operate directly on rules and worlds. In this chapter we will introduce the system Tartan Worlds as implemented on the Apple Macintosh.
series ACADIA
email
last changed 2022/06/07 07:57

_id aa78
authors Bayazit, Nigan
year 1992
title Requirements of an Expert System for Design Studios
doi https://doi.org/10.52842/conf.ecaade.1992.187
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 187-194
summary The goal of this paper is to study problems of the transition from traditional architectural studio teaching to CAAD studio teaching which requires a CAAD expert system as studio tutor, and to study the behavior of the student in this new environment. The differences between the traditional and computerized studio teaching and the experiences in this field are explained referring to the requirements for designing time in relation to the expertise of the student in the application of a CAD program. Learning styles and the process of design in computerized and non-computerized studio teaching are discussed. Design studio requirements of the students in traditional studio environment while doing design works are clarified depending on the results of an empirical study which explained the relations between the tutor and the student while they were doing studio critiques. Main complaints of the students raised in the empirical study were the lack of data in the specific design problem area, difficulties of realization of ideas and thoughts, not knowing the starting point of design, having no information about the references to be used for the specific design task, having difficulties in the application of presentation techniques. In the concluding parts of the paper are discussed the different styles of teaching and their relation to the CAAD environment, the transformation of the instructional programs for the new design environment, the future expectations from the CAAD programs, properties of the new teaching environment and the roles of the expert systems in design studio education.

keywords CAAD Education, Expert System, Architectural Design Studio, Knowledge Acquisition
series eCAADe
email
last changed 2022/06/07 07:54

_id cef3
authors Bridges, Alan H.
year 1992
title Computing and Problem Based Learning at Delft University of Technology Faculty of Architecture
doi https://doi.org/10.52842/conf.ecaade.1992.289
source CAAD Instruction: The New Teaching of an Architect? [eCAADe Conference Proceedings] Barcelona (Spain) 12-14 November 1992, pp. 289-294
summary Delft University of Technology, founded in 1842, is the oldest and largest technical university in the Netherlands. It provides education for more than 13,000 students in fifteen main subject areas. The Faculty of Architecture, Housing, Urban Design and Planning is one of the largest faculties of the DUT with some 2000 students and over 500 staff members. The course of study takes four academic years: a first year (Propaedeuse) and a further three years (Doctoraal) leading to the "ingenieur" qualification. The basic course material is delivered in the first two years and is taken by all students. The third and fourth years consist of a smaller number of compulsory subjects in each of the department's specialist areas together with a wide range of option choices. The five main subject areas the students may choose from for their specialisation are Architecture, Building and Project Management, Building Technology, Urban Design and Planning, and Housing.

The curriculum of the Faculty has been radically revised over the last two years and is now based on the concept of "Problem-Based Learning". The subject matter taught is divided thematically into specific issues that are taught in six week blocks. The vehicles for these blocks are specially selected and adapted case studies prepared by teams of staff members. These provide a focus for integrating specialist subjects around a studio based design theme. In the case of second year this studio is largely computer-based: many drawings are produced by computer and several specially written computer applications are used in association with the specialist inputs.

This paper describes the "block structure" used in second year, giving examples of the special computer programs used, but also raises a number of broader educational issues. Introduction of the block system arose as a method of curriculum integration in response to difficulties emerging from the independent functioning of strong discipline areas in the traditional work groups. The need for a greater level of selfdirected learning was recognised as opposed to the "passive information model" of student learning in which the students are seen as empty vessels to be filled with knowledge - which they are then usually unable to apply in design related contexts in the studio. Furthermore, the value of electives had been questioned: whilst enabling some diversity of choice, they may also be seen as diverting attention and resources from the real problems of teaching architecture.

series eCAADe
email
last changed 2022/06/07 07:54

_id cf5c
authors Carpenter, B.
year 1992
title The logic of typed feature structures with applications to unification grammars, logic programs and constraint resolution
source Cambridge Tracts in Theoretical Computer Science, Cambridge University Press
summary This book develops the theory of typed feature structures, a new form of data structure that generalizes both the first-order terms of logic programs and feature-structures of unification-based grammars to include inheritance, typing, inequality, cycles and intensionality. It presents a synthesis of many existing ideas into a uniform framework, which serves as a logical foundation for grammars, logic programming and constraint-based reasoning systems. Throughout the text, a logical perspective is adopted that employs an attribute-value description language along with complete equational axiomatizations of the various systems of feature structures. Efficiency concerns are discussed and complexity and representability results are provided. The application of feature structures to phrase structure grammars is described and completeness results are shown for standard evaluation strategies. Definite clause logic programs are treated as a special case of phrase structure grammars. Constraint systems are introduced and an enumeration technique is given for solving arbitrary attribute-value logic constraints. This book with its innovative approach to data structures will be essential reading for researchers in computational linguistics, logic programming and knowledge representation. Its self-contained presentation makes it flexible enough to serve as both a research tool and a textbook.
series other
last changed 2003/04/23 15:14

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 12HOMELOGIN (you are user _anon_969589 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002