CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 4858

_id 63d0
authors Carrara, Gianfranco and Novembri, Gabriele
year 1986
title Constraint-bounded design search
source Computer-Aided Architectural Design Futures [CAAD Futures Conference Proceedings / ISBN 0-408-05300-3] Delft (The Netherlands), 18-19 September 1985, pp. 146-157
summary The design process requires continual checking of the consistency of design choices against given sets of goals that have been fulfilled. Such a check is generally performed by comparing abstract representations of design goals with these of the sought real building objects (RBO) resulting from complex intellectual activities closely related to the designer's culture and to the environment in which he operates. In this chapter we define a possible formalization of such representations concerning the goals and the RBO that are usually considered in the architectural design process by our culture in our environment. The representation of design goals is performed by expressing their objective aspects (requirements) and by defining their allowable values (performance specifications). The resulting system of requirements defines the set of allowable solutions and infers an abstract representation of the sought building objects (BO) that consists of the set of characteristics (attributes and relations) which are considered relevant to represent the particular kind of RBO with respect to the consistency check with design goals. The values related to such characteristics define the performances of the RBO while their set establishes its behaviour. Generally speaking, there is no single real object corresponding to an abstract representation but the whole class of the RBO that are equivalent with respect to the values assumed by the considered characteristics. The more we increase the number of these, as well as their specifications, the smaller the class becomes until it coincides with a single real object - given that the assessed specifications be fully consistent. On the other hand, the corresponding representation evolves to the total prefiguration of the RBO. It is not therefore possible to completely define a BO representation in advance since this is inferred by the considered goals and is itself a result of the design process. What can only be established in advance is that any set of characteristics assumed to represent any RBO consists of hierarchic, topological, geometrical and functional relations among the parts of the object at any level of aggregation (from components to space units, to building units, to the whole building) that we define representation structure (RS). Consequently the RS may be thought as the elementary structures that, by superposition and interaction, set up the abstract representation that best fit with design goals.
series CAAD Futures
last changed 1999/04/03 17:58

_id a18b
authors Samet, Hanan and Webber, Robert E.
year 1985
title Storing a Collection of Polygons Using Quadtrees
source ACM Transactions on Graphics July, 1985. vol. 4: pp. 182-222 : some ill. includes bibliography.
summary An adaptation of the quadtree data structure that represents polygonal maps (i.e., collections of polygons, possibly containing holes) is described in a manner that is also useful for the manipulation of arbitrary collections of straight line segments. The goal is to store these maps without the loss of information that results from digitization, and to obtain a worst-case execution time that is not overly sensitive to the positioning of the map. Regular decomposition variant of the region quadtree is usedÔ h)0*0*0*°° ÔŒ to organize the vertices and edges of the maps. A number of related data organizations are proposed in an iterative manner until a method is obtained that meets the stated goals. The result is termed a PM (Polygonal Map) quadtree and is based on a regular decomposition Point Space quadtree (PS quadtree) that stores additional information about the edges at its terminal nodes. Algorithms are given for inserting and deleting line segments from a PM quadtree. Use of the PM quadtree to perform point location, dynamic line insertion, and map overlay is discussed. An empirical comparison of the PM quadtree with other quadtree-based representations for polygonal maps is also provided
keywords data structures, quadtree, polygons, representation, point inclusion, algorithms
series CADline
last changed 2003/06/02 10:24

_id cc15
authors Ansaldi, Silvia, De Floriani, Leila and Falcidieno, Bianca
year 1985
title Geometric Modeling of Solid Objects by Using a Face Adjacency Graph Representation
source SIGGRAPH '85 Conference Proceedings. July, 1985. vol. 19 ; no. 3: pp. 131-139 : ill. includes bibliography
summary A relational graph structure based on a boundary representation of solid objects is described. In this structure, called Face Adjacency Graph, nodes represent object faces, whereas edges and vertices are encoded into arcs and hyperarcs. Based on the face adjacency graph, the authors define a set of primitive face-oriented Euler operators, and a set of macro operators for face manipulation, which allow a compact definition and an efficient updating of solid objects. The authors briefly describe a hierarchical graph structure based on the face adjacency graph, which provides a representation of an object at different levels of detail. Thus it is consistent with the stepwise refinement process through which the object description is produced
keywords geometric modeling, graphs, objects, representation, data structures,B-rep, solid modeling, Euler operators
series CADline
last changed 2003/06/02 10:24

_id acfe
authors Archea, John
year 1985
title Architecture's Unique Position Among the Disciplines : Puzzle-Making vs. Problem Solving
source CRIT XV, The Architectural Student Journal. Summer, 1985. pp. 20-22
summary Most disciplines involved in the building process, i.e., programmers, space planners, and engineers work in what may be described as a problem solving mode. They state desired effects as explicit performance criteria before they initiate a decision process and test alternative solutions against those criteria until a fit is attained which falls within known probabilities of success. Architects, however are not problem solvers and they are not seeking explicit information when they design how buildings work. Architects are puzzle- makers, They are primarily concerned with unique design concepts. It is through the act of designing, or puzzle- making, that the architect learn what they want to accomplish and how. With regard to the making of buildings, places or experiences, the architect is a puzzle-maker surrounded by a group of problem solvers who address separate pieces of the puzzle
keywords puzzle making, design process, problem solving, architecture
series CADline
last changed 1999/02/12 15:07

_id 8ae8
authors Ayala, D., P. Brunet and Juan (et al)
year 1985
title Object Representation by Means of Nominimal Division Quadtrees and Octrees
source ACM Transactions on Graphics. January, 1985. vol. 4: pp. 41-59 : ill. includes bibliography
summary Quadtree representation of two-dimensional objects is performed with a tree that describes the recursive subdivision of the more complex parts of a picture until the desired resolution is reached. At the end, all the leaves of the tree are square cells that lie completely inside or outside the object. There are two great disadvantages in the use of quadtrees as a representation scheme for objects in geometric modeling system: The amount of memory required for polygonal objects is too great, and it is difficult to recompute the boundary representation of the object after some Boolean operations have been performed. In the present paper a new class of quadtrees, in which nodes may contain zero or one edge, is introduced. By using these quadtrees, storage requirements are reduced and it is possible to obtain the exact backward conversion to boundary representation. Algorithms for the generation of the quadtree, boolean operation, and recomputation of the boundary representation are presented, and their complexities in time and space are discussed. Three- dimensional algorithms working on octrees are also presented. Their use in the geometric modeling of three-dimensional polyhedral objects is discussed
keywords geometric modeling, algorithms, octree, quadtree, curves, curved surfaces, boolean operations
series CADline
last changed 2003/06/02 13:58

_id c50a
authors Bartschi, Martin
year 1985
title An Overview of Information Retrieval Subjects
source IEEE Computer. May, 1985. vol. 18: pp. 67-84 : ill. includes bibliography
summary The aim of an information retrieval system is to find information items relevant to an information need. As relevance is a kind of similarity relation between the concepts represented by the information item and those represented by the formulation of the information need, it is not astonishing to discover that the class of possible query forms -formulations of the information needs - is the same as the class of possible representations of information items. This article overviews current research problems in information structure and query evaluation
keywords database, information, queries, systems
series CADline
last changed 1999/02/12 15:07

_id ddssar0206
id ddssar0206
authors Bax, M.F.Th. and Trum, H.M.G.J.
year 2002
title Faculties of Architecture
source Timmermans, Harry (Ed.), Sixth Design and Decision Support Systems in Architecture and Urban Planning - Part one: Architecture Proceedings Avegoor, the Netherlands), 2002
summary In order to be inscribed in the European Architect’s register the study program leading to the diploma ‘Architect’ has to meet the criteria of the EC Architect’s Directive (1985). The criteria are enumerated in 11 principles of Article 3 of the Directive. The Advisory Committee, established by the European Council got the task to examine such diplomas in the case some doubts are raised by other Member States. To carry out this task a matrix was designed, as an independent interpreting framework that mediates between the principles of Article 3 and the actual study program of a faculty. Such a tool was needed because of inconsistencies in the list of principles, differences between linguistic versions ofthe Directive, and quantification problems with time, devoted to the principles in the study programs. The core of the matrix, its headings, is a categorisation of the principles on a higher level of abstractionin the form of a taxonomy of domains and corresponding concepts. Filling in the matrix means that each study element of the study programs is analysed according to their content in terms of domains; thesummation of study time devoted to the various domains results in a so-called ‘profile of a faculty’. Judgement of that profile takes place by committee of peers. The domains of the taxonomy are intrinsically the same as the concepts and categories, needed for the description of an architectural design object: the faculties of architecture. This correspondence relates the taxonomy to the field of design theory and philosophy. The taxonomy is an application of Domain theory. This theory,developed by the authors since 1977, takes as a view that the architectural object only can be described fully as an integration of all types of domains. The theory supports the idea of a participatory andinterdisciplinary approach to design, which proved to be awarding both from a scientific and a social point of view. All types of domains have in common that they are measured in three dimensions: form, function and process, connecting the material aspects of the object with its social and proceduralaspects. In the taxonomy the function dimension is emphasised. It will be argued in the paper that the taxonomy is a categorisation following the pragmatistic philosophy of Charles Sanders Peirce. It will bedemonstrated as well that the taxonomy is easy to handle by giving examples of its application in various countries in the last 5 years. The taxonomy proved to be an adequate tool for judgement ofstudy programs and their subsequent improvement, as constituted by the faculties of a Faculty of Architecture. The matrix is described as the result of theoretical reflection and practical application of a matrix, already in use since 1995. The major improvement of the matrix is its direct connection with Peirce’s universal categories and the self-explanatory character of its structure. The connection with Peirce’s categories gave the matrix a more universal character, which enables application in other fieldswhere the term ‘architecture’ is used as a metaphor for artefacts.
series DDSS
last changed 2003/11/21 15:16

_id ddss9408
id ddss9408
authors Bax, Thijs and Trum, Henk
year 1994
title A Taxonomy of Architecture: Core of a Theory of Design
source Second Design and Decision Support Systems in Architecture & Urban Planning (Vaals, the Netherlands), August 15-19, 1994
summary The authors developed a taxonomy of concepts in architectural design. It was accepted by the Advisory Committee for education in the field of architecture, a committee advising the European Commission and Member States, as a reference for their task to harmonize architectural education in Europe. The taxonomy is based on Domain theory, a theory developed by the authors, based on General Systems Theory and the notion of structure according to French Structuralism, takes a participatory viewpoint for the integration of knowledge and interests by parties in the architectural design process. The paper discusses recent developments of the taxonomy, firstly as a result of a confrontation with similar endeavours to structure the field of architectural design, secondly as a result of applications of education and architectural design practice, and thirdly as a result of theapplication of some views derived from the philosophical work from Charles Benjamin Peirce. Developments concern the structural form of the taxonomy comprising basic concepts and levelbound scale concepts, and the specification of the content of the fields which these concepts represent. The confrontation with similar endeavours concerns mainly the work of an ARCUK workingparty, chaired by Tom Marcus, based on the European Directive from 1985. The application concerns experiences with a taxonomy-based enquiry in order to represent the profile of educational programmes of schools and faculties of architecture in Europe in qualitative and quantitative terms. This enquiry was carried out in order to achieve a basis for comparison and judgement, and a basis for future guidelines including quantitative aspects. Views of Peirce, more specifically his views on triarchy as a way of ordering and structuring processes of thinking,provide keys for a re-definition of concepts as building stones of the taxonomy in terms of the form-function-process-triad, which strengthens the coherence of the taxonomy, allowing for a more regular representation in the form of a hierarchical ordered matrix.
series DDSS
last changed 2003/08/07 16:36

_id 66b3
authors Bollinger, Elizabeth
year 1985
title Integrating CADD into the AEC Process - A Case Study
doi https://doi.org/10.52842/conf.acadia.1985.013
source ACADIA Workshop ‘85 [ACADIA Conference Proceedings] Tempe (Arizona / USA) 2-3 November 1985, pp. 13-24
summary A research grant was awarded to the Graduate School of Architecture at the University of Houston by Nash Phillips/Copus, a large homebuilding corporation, to study the integration of computer aided design into the entire building process. A computer aided design system had been utilized by the firm's department of architecture and planning for several months. A team of University faculty and graduate students studied the organization of the firm with respect to functions that could be automated. Its determination was that by utilizing an integrated data base, with information to be extracted from the computer generated drawings, the entire process of bidding and building a structure could be made more efficient and cost effective. The research team developed a system in which cost estimating could be done directly from the drawings. As drawings were modified, new reports could be automatically generated. More design solutions could be studied from the impact of cost as well as aesthetics. Additionally, once plans were drawn, a program written by students would automatically generate elevations of wall panels to be sent to the construction department for its use, and which would also generate material reports. The team also studied techniques of computer modelling for usage by the architectural planning department in client presentations.
series ACADIA
email EBollinger@uh.edu
last changed 2022/06/07 07:54

_id avocaad_2001_02
id avocaad_2001_02
authors Cheng-Yuan Lin, Yu-Tung Liu
year 2001
title A digital Procedure of Building Construction: A practical project
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary In earlier times in which computers have not yet been developed well, there has been some researches regarding representation using conventional media (Gombrich, 1960; Arnheim, 1970). For ancient architects, the design process was described abstractly by text (Hewitt, 1985; Cable, 1983); the process evolved from unselfconscious to conscious ways (Alexander, 1964). Till the appearance of 2D drawings, these drawings could only express abstract visual thinking and visually conceptualized vocabulary (Goldschmidt, 1999). Then with the massive use of physical models in the Renaissance, the form and space of architecture was given better precision (Millon, 1994). Researches continued their attempts to identify the nature of different design tools (Eastman and Fereshe, 1994). Simon (1981) figured out that human increasingly relies on other specialists, computational agents, and materials referred to augment their cognitive abilities. This discourse was verified by recent research on conception of design and the expression using digital technologies (McCullough, 1996; Perez-Gomez and Pelletier, 1997). While other design tools did not change as much as representation (Panofsky, 1991; Koch, 1997), the involvement of computers in conventional architecture design arouses a new design thinking of digital architecture (Liu, 1996; Krawczyk, 1997; Murray, 1997; Wertheim, 1999). The notion of the link between ideas and media is emphasized throughout various fields, such as architectural education (Radford, 2000), Internet, and restoration of historical architecture (Potier et al., 2000). Information technology is also an important tool for civil engineering projects (Choi and Ibbs, 1989). Compared with conventional design media, computers avoid some errors in the process (Zaera, 1997). However, most of the application of computers to construction is restricted to simulations in building process (Halpin, 1990). It is worth studying how to employ computer technology meaningfully to bring significant changes to concept stage during the process of building construction (Madazo, 2000; Dave, 2000) and communication (Haymaker, 2000).In architectural design, concept design was achieved through drawings and models (Mitchell, 1997), while the working drawings and even shop drawings were brewed and communicated through drawings only. However, the most effective method of shaping building elements is to build models by computer (Madrazo, 1999). With the trend of 3D visualization (Johnson and Clayton, 1998) and the difference of designing between the physical environment and virtual environment (Maher et al. 2000), we intend to study the possibilities of using digital models, in addition to drawings, as a critical media in the conceptual stage of building construction process in the near future (just as the critical role that physical models played in early design process in the Renaissance). This research is combined with two practical building projects, following the progress of construction by using digital models and animations to simulate the structural layouts of the projects. We also tried to solve the complicated and even conflicting problems in the detail and piping design process through an easily accessible and precise interface. An attempt was made to delineate the hierarchy of the elements in a single structural and constructional system, and the corresponding relations among the systems. Since building construction is often complicated and even conflicting, precision needed to complete the projects can not be based merely on 2D drawings with some imagination. The purpose of this paper is to describe all the related elements according to precision and correctness, to discuss every possibility of different thinking in design of electric-mechanical engineering, to receive feedback from the construction projects in the real world, and to compare the digital models with conventional drawings.Through the application of this research, the subtle relations between the conventional drawings and digital models can be used in the area of building construction. Moreover, a theoretical model and standard process is proposed by using conventional drawings, digital models and physical buildings. By introducing the intervention of digital media in design process of working drawings and shop drawings, there is an opportune chance to use the digital media as a prominent design tool. This study extends the use of digital model and animation from design process to construction process. However, the entire construction process involves various details and exceptions, which are not discussed in this paper. These limitations should be explored in future studies.
series AVOCAAD
email aleppo@cc.nctu.edu.tw
last changed 2005/09/09 10:48

_id 298e
authors Dave, Bharat and Woodbury, Robert
year 1990
title Computer Modeling: A First Course in Design Computing
source The Electronic Design Studio: Architectural Knowledge and Media in the Computer Era [CAAD Futures ‘89 Conference Proceedings / ISBN 0-262-13254-0] Cambridge (Massachusetts / USA), 1989, pp. 61-76
summary Computation in design has long been a focus in our department. In recent years our faculty has paid particular attention to the use of computation in professional architectural education. The result is a shared vision of computers in the curriculum [Woodbury 1985] and a set of courses, some with considerable historyland others just now being initiated. We (Dave and Woodbury) have jointly developed and at various times over the last seven years have taught Computer Modeling, the most introductory of these courses. This is a required course for all the incoming freshmen students in the department. In this paper we describe Computer Modeling: its context, the issues and topics it addresses, the tasks it requires of students, and the questions and opportunities that it raises. Computer Modeling is a course about concepts, about ways of explicitly understanding design and its relation to computation. Procedural skills and algorithmic problem solving techniques are given only secondary emphasis. In essential terms, the course is about models, of design processes, of designed objects, of computation and of computational design. Its lessons are intended to communicate a structure of such models to students and through this structure to demonstrate a relationship between computation and design. It is hoped that this structure can be used as a framework, around which students can continue to develop an understanding of computers in design.
series CAAD Futures
email b.dave@unimelb.edu.au
last changed 2003/05/16 20:58

_id 0faa
authors Duelund Mortensen, Peder
year 1991
title THE FULL-SCALE MODEL WORKSHOP
source Proceedings of the 3rd European Full-Scale Modelling Conference / ISBN 91-7740044-5 / Lund (Sweden) 13-16 September 1990, pp. 10-11
summary The workshop is an institution, available for use by the public and established at the Laboratory of Housing in the Art Academy's school of Architecture for a 3 year trial period beginning April 1985. This resumé contains brief descriptions of a variety of representative model projects and an overview of all projects carried out so far, including the pilot projects from 1983 and planned projects to and including January 1987. The Full Scale Model Workshop builds full size models of buildings, rooms and parts of buildings. The purpose of the Full Scale Model Workshop is to promote communication among building's users. The workshop is a tool in an attempt to build bridges between theory and practice in research, experimentation and communication of research results. New ideas and experiments of various sorts can be tried out cheaply, quickly and efficiently through the building of full scale models. Changes can be done on the spot as a planned part of the project and on the basis of ideas and experiments achieved through the model work itself. Buildings and their space can thus be communicated directly to all involved persons, regardless of technical background or training in evaluation of building projects.
keywords Full-scale Modeling, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 15:23

_id acadia06_068
id acadia06_068
authors Elys, John
year 2006
title Digital Ornament
doi https://doi.org/10.52842/conf.acadia.2006.068
source Synthetic Landscapes [Proceedings of the 25th Annual Conference of the Association for Computer-Aided Design in Architecture] pp. 68-78
summary Gaming software has a history of fostering development of economical and creative methods to deal with hardware limitations. Traditionally the visual representation of gaming software has been a poor offspring of high-end visualization. In a twist of irony, this paper proposes that game production software leads the way into a new era of physical digital ornament. The toolbox of the rendering engine evolved rapidly between 1974-1985 and it is still today, 20 years later the main component of all visualization programs. The development of the bump map is of particular interest; its evolution into a physical displacement map provides untold opportunities of the appropriation of the 2D image to a physical 3D object.To expose the creative potential of the displacement map, a wide scope of existing displacement usage has been identified: Top2maya is a scientific appropriation, Caruso St John Architects an architectural precedent and Tord Boonje’s use of 2D digital pattern provides us with an artistic production precedent. Current gaming technologies give us an indication of how the resolution of displacement is set to enter an unprecedented level of geometric detail. As modernity was inspired by the machine age, we should be led by current technological advancement and appropriate its usage. It is about a move away from the simplification of structure and form to one that deals with the real possibilities of expanding the dialogue of surface topology. Digital Ornament is a kinetic process rather than static, its intentions lie in returning the choice of bespoke materials back to the Architect, Designer and Artist.
series ACADIA
email elysjohn@hotmail.com
last changed 2022/06/07 07:55

_id ga0024
id ga0024
authors Ferrara, Paolo and Foglia, Gabriele
year 2000
title TEAnO or the computer assisted generation of manufactured aesthetic goods seen as a constrained flux of technological unconsciousness
source International Conference on Generative Art
summary TEAnO (Telematica, Elettronica, Analisi nell'Opificio) was born in Florence, in 1991, at the age of 8, being the direct consequence of years of attempts by a group of computer science professionals to use the digital computers technology to find a sustainable match among creation, generation (or re-creation) and recreation, the three basic keywords underlying the concept of “Littérature potentielle” deployed by Oulipo in France and Oplepo in Italy (see “La Littérature potentielle (Créations Re-créations Récréations) published in France by Gallimard in 1973). During the last decade, TEAnO has been involving in the generation of “artistic goods” in aesthetic domains such as literature, music, theatre and painting. In all those artefacts in the computer plays a twofold role: it is often a tool to generate the good (e.g. an editor to compose palindrome sonnets of to generate antonymic music) and, sometimes it is the medium that makes the fruition of the good possible (e.g. the generator of passages of definition literature). In that sense such artefacts can actually be considered as “manufactured” goods. A great part of such creation and re-creation work has been based upon a rather small number of generation constraints borrowed from Oulipo, deeply stressed by the use of the digital computer massive combinatory power: S+n, edge extraction, phonetic manipulation, re-writing of well known masterpieces, random generation of plots, etc. Regardless this apparently simple underlying generation mechanisms, the systematic use of computer based tools, as weel the analysis of the produced results, has been the way to highlight two findings which can significantly affect the practice of computer based generation of aesthetic goods: ? the deep structure of an aesthetic work persists even through the more “desctructive” manipulations, (such as the antonymic transformation of the melody and lyrics of a music work) and become evident as a sort of profound, earliest and distinctive constraint; ? the intensive flux of computer generated “raw” material seems to confirm and to bring to our attention the existence of what Walter Benjamin indicated as the different way in which the nature talk to a camera and to our eye, and Franco Vaccari called “technological unconsciousness”. Essential references R. Campagnoli, Y. Hersant, “Oulipo La letteratura potenziale (Creazioni Ri-creazioni Ricreazioni)”, 1985 R. Campagnoli “Oupiliana”, 1995 TEAnO, “Quaderno n. 2 Antologia di letteratura potenziale”, 1996 W. Benjiamin, “Das Kunstwerk im Zeitalter seiner technischen Reprodizierbarkeit”, 1936 F. Vaccari, “Fotografia e inconscio tecnologico”, 1994
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 78ca
authors Friedland, P. (Ed.)
year 1985
title Special Section on Architectures for Knowledge-Based Systems
source CACM (28), 9, September
summary A fundamental shift in the preferred approach to building applied artificial intelligence (AI) systems has taken place since the late 1960s. Previous work focused on the construction of general-purpose intelligent systems; the emphasis was on powerful inference methods that could function efficiently even when the available domain-specific knowledge was relatively meager. Today the emphasis is on the role of specific and detailed knowledge, rather than on reasoning methods.The first successful application of this method, which goes by the name of knowledge-based or expert-system research, was the DENDRAL program at Stanford, a long-term collaboration between chemists and computer scientists for automating the determination of molecular structure from empirical formulas and mass spectral data. The key idea is that knowledge is power, for experts, be they human or machine, are often those who know more facts and heuristics about a domain than lesser problem solvers. The task of building an expert system, therefore, is predominantly one of teaching" a system enough of these facts and heuristics to enable it to perform competently in a particular problem-solving context. Such a collection of facts and heuristics is commonly called a knowledge base. Knowledge-based systems are still dependent on inference methods that perform reasoning on the knowledge base, but experience has shown that simple inference methods like generate and test, backward-chaining, and forward-chaining are very effective in a wide variety of problem domains when they are coupled with powerful knowledge bases. If this methodology remains preeminent, then the task of constructing knowledge bases becomes the rate-limiting factor in expert-system development. Indeed, a major portion of the applied AI research in the last decade has been directed at developing techniques and tools for knowledge representation. We are now in the third generation of such efforts. The first generation was marked by the development of enhanced AI languages like Interlisp and PROLOG. The second generation saw the development of knowledge representation tools at AI research institutions; Stanford, for instance, produced EMYCIN, The Unit System, and MRS. The third generation is now producing fully supported commercial tools like KEE and S.1. Each generation has seen a substantial decrease in the amount of time needed to build significant expert systems. Ten years ago prototype systems commonly took on the order of two years to show proof of concept; today such systems are routinely built in a few months. Three basic methodologies-frames, rules, and logic-have emerged to support the complex task of storing human knowledge in an expert system. Each of the articles in this Special Section describes and illustrates one of these methodologies. "The Role of Frame-Based Representation in Reasoning," by Richard Fikes and Tom Kehler, describes an object-centered view of knowledge representation, whereby all knowldge is partitioned into discrete structures (frames) having individual properties (slots). Frames can be used to represent broad concepts, classes of objects, or individual instances or components of objects. They are joined together in an inheritance hierarchy that provides for the transmission of common properties among the frames without multiple specification of those properties. The authors use the KEE knowledge representation and manipulation tool to illustrate the characteristics of frame-based representation for a variety of domain examples. They also show how frame-based systems can be used to incorporate a range of inference methods common to both logic and rule-based systems.""Rule-Based Systems," by Frederick Hayes-Roth, chronicles the history and describes the implementation of production rules as a framework for knowledge representation. In essence, production rules use IF conditions THEN conclusions and IF conditions THEN actions structures to construct a knowledge base. The autor catalogs a wide range of applications for which this methodology has proved natural and (at least partially) successful for replicating intelligent behavior. The article also surveys some already-available computational tools for facilitating the construction of rule-based knowledge bases and discusses the inference methods (particularly backward- and forward-chaining) that are provided as part of these tools. The article concludes with a consideration of the future improvement and expansion of such tools.The third article, "Logic Programming, " by Michael Genesereth and Matthew Ginsberg, provides a tutorial introduction to the formal method of programming by description in the predicate calculus. Unlike traditional programming, which emphasizes how computations are to be performed, logic programming focuses on the what of objects and their behavior. The article illustrates the ease with which incremental additions can be made to a logic-oriented knowledge base, as well as the automatic facilities for inference (through theorem proving) and explanation that result from such formal descriptions. A practical example of diagnosis of digital device malfunctions is used to show how significantand complex problems can be represented in the formalism.A note to the reader who may infer that the AI community is being split into competing camps by these three methodologies: Although each provides advantages in certain specific domains (logic where the domain can be readily axiomatized and where complete causal models are available, rules where most of the knowledge can be conveniently expressed as experiential heuristics, and frames where complex structural descriptions are necessary to adequately describe the domain), the current view is one of synthesis rather than exclusivity. Both logic and rule-based systems commonly incorporate frame-like structures to facilitate the representation of large amounts of factual information, and frame-based systems like KEE allow both production rules and predicate calculus statements to be stored within and activated from frames to do inference. The next generation of knowledge representation tools may even help users to select appropriate methodologies for each particular class of knowledge, and then automatically integrate the various methodologies so selected into a consistent framework for knowledge. "
series journal paper
last changed 2003/04/23 15:14

_id 027b
authors Griffiths, J.G.
year 1985
title Table-Driven Algorithms for Generating Space-Filling Curves
source Computer Aided Design. January/ February, 1985. vol. 17: pp. 37-41 : ill. includes bibliography
summary A simple general method for constructing space-filling curves is presented, based on the use of tables. It is shown how the use of Hilbert's curve can enhance the performance of Warnock's algorithm. A procedure is given which generates Hilbert curves or Sierpinski curves. A second procedure is given which generates Warnock's windows in Hilbert order
keywords computer graphics, rendering, algorithms, curves, representation, display
series CADline
last changed 2003/06/02 13:58

_id 76ce
authors Grimson, W.
year 1985
title Computational Experiments with a Feature Based Stereo Algorithm
source IEEE Trans. Pattern Anal. Machine Intell., Vol. PAMI-7, No. 1
summary Computational models of the human stereo system' can provide insight into general information processing constraints that apply to any stereo system, either artificial or biological. In 1977, Marr and Poggio proposed one such computational model, that was characterized as matching certain feature points in difference-of-Gaussian filtered images, and using the information obtained by matching coarser resolution representations to restrict the search'space for matching finer resolution representations. An implementation of the algorithm and'its testing on a range of images was reported in 1980. Since then a number of psychophysical experiments have suggested possible refinements to the model and modifications to the algorithm. As well, recent computational experiments applying the algorithm to a variety of natural images, especially aerial photographs, have led to a number of modifications. In this article, we present a version of the Marr-Poggio-Gfimson algorithm that embodies these modifications and illustrate its performance on a series of natural images.
series journal paper
last changed 2003/04/23 15:14

_id 07c6
authors Kalay, Y.E., Harfmann, A.C. and Swerdloff, L.M.
year 1985
title ALEX: A Knowledge-Based Architectural Design System
doi https://doi.org/10.52842/conf.acadia.1985.096
source ACADIA Workshop ‘85 [ACADIA Conference Proceedings] Tempe (Arizona / USA) 2-3 November 1985, pp. 96-108
summary A methodology for the development of a knowledge-based computer-aided design system and its experimental application in the domain of single family house design are presented.

The methodology involves integrating within a unified design environment, tools and techniques that have been independently developed in various disciplines (including knowledge representation, information management, geometric modeling, human,machine interface, and architectural design). By assuming the role of active design partners, the resulting systems are expected to increase the productivity of designers, improve the quality of their products, and reduce cost and lead time of the design process as a whole.

ALEX (Architecture Learning Expert), a particular application of this methodology, is a prototype knowledge-based CAD system in the domain of single family house design. It employs user-interactive, goal directed heuristic search strategies in a solution space that consists of a network of objects. Message-based change propagation techniques, guided by domain-specific knowledge, are used to ensure database integrity and well-formedness.

The significance of the methodology and its application is threefold: it furthers our knowledge of the architectural design process, explores the utilization of knowledge engineering methods in design, and serves as a prototype for developing the next generation of computer-aided architectural design systems.

series ACADIA
email kalay@socrates.berkeley.edu
last changed 2022/06/07 07:52

_id e1a8
authors Kellogg, Richard E.
year 1985
title CAD-Spreadsheet Linkages for Design and Analysis
doi https://doi.org/10.52842/conf.acadia.1985.109
source ACADIA Workshop ‘85 [ACADIA Conference Proceedings] Tempe (Arizona / USA) 2-3 November 1985, pp. 109-118
summary This paper reports on two systems under development which link a CAD system with a spreadsheet. The first extracts areas and R-values from a special AutoCAD drawing and processes the information in a Lotus 1-2-3 spreadsheet to obtain total heatloss for a building. The second is a prototype expert system which uses space labels from an AutoCAD "bubble-diagram" to print lists of design recommendations extracted from a Lotus 1-2-3 data-base. These methods emphasize drawing as the primary design activity, while providing immediate factual feedback about the design proposal.

series ACADIA
email rkellogg@uark.edu
last changed 2022/06/07 07:52

_id a920
authors Kulcke, Richard
year 1989
title CAAD in the Architectural Education of the Fachhochschulen in the Federal Republic of Germany
doi https://doi.org/10.52842/conf.ecaade.1989.x.w7a
source CAAD: Education - Research and Practice [eCAADe Conference Proceedings / ISBN 87-982875-2-4] Aarhus (Denmark) 21-23 September 1989, pp. 4.3.1
summary For over 10 years the author has been a teacher in the field of "computer application in architecture" at the Fachhochschule. Since 1985 he regularly has been taking part in the conferences of A.I.I.D.A. (Arbeitskreis INFORMATIK IN DER ARCHlTEKTENAUSBILDUNG). All the faculties of architecture at the Fachhochschulen (about 10) can send their representatives of CAAD to the conferences. A.I.I.D.A. has been having 2 conferences a year since 1985. At the last conference in Wiesbaden a paper with statements of A.I.I.D.A. for the further education in CAAD was finished. The author presents and explains this paper. On the other hand he shows the actual education program of CAAD of his faculty. The education in CAAD started in 1972 with basic information without practical elements. Now the practical work with the workstation is talking most of the time . The computer application is available for subjects like Building Economics, Building and Structure Design and others. With his assistant the author developed programs of the field of Building Economics. In 1986 he started introduce CAD with AutoCAD in the education program. Now also other colleagues start to integrate CAAD into their subjects.

series eCAADe
last changed 2022/06/07 07:50

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 242HOMELOGIN (you are user _anon_198282 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002