CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 77

_id 48a2
authors Bergstrom, Axel Lennart
year 1982
title Computer Graphics in Community Planning
source June, 1982. 3 p
summary The digitized map of Sweden uses a technique for graphical presentation in color based on an automatic drawing machine called the 'color ink jet plotter.' A short description of this project is given
keywords mapping, computer graphics
series CADline
last changed 1999/02/12 15:07

_id 6094
authors Blinn, J.I.
year 1982
title A Generalization of Algebraic Surface Drawing
source ACM Transaction on Graphics, vol. 1, no. 3, pp. 235-256, 1982
summary The technology of creating realistic and visually interesting images of three- dimensional shapes is advancing on many fronts. One such front is the develop- ment of algorithms for drawing curved surfaces directly from their mathematical definitions rather than by dividing them into large numbers of polygons. Two classes of surfaces which have received attention are the quadric and the bivariate parametric surfaces. Bivariate parametric surfaces are generated by three func- tions of two variables (most popularly polynomials), as the variables take on different values. Algorithms dealing with such surfaces are due to Catmull; Lane, Carpenter, Whitted and Blinn; and Clark.
series journal paper
last changed 2003/11/21 15:16

_id 0dfa
authors Giuse, Dario
year 1982
title DP - Command Set
source 24, [22] p. : ill. July, 1982. includes bibliography
summary This is a short description of the command set of the Drawing Program (DP), a general-purpose graphics editor that runs on a personal computer. This paper contains general notions about the program, the complete description of the commands and a set of advanced techniques for DP users
keywords software, programming, drafting, systems
series CADline
last changed 1999/02/12 15:08

_id 5509
authors Koutamanis, Alexandros
year 1990
title Development of a computerized handbook of architectural plans
source Delft University of Technology
summary The dissertation investigates an approach to the development of visual / spatial computer representations for architectural purposes through the development of the computerized handbook of architectural plans (chap), a knowledge-based computer system capable of recognizing the metric properties of architectural plans. This investigation can be summarized as an introduction of computer vision to the computerization of architectural representations: chap represents an attempt to automate recognition of the most essential among conventional architectural drawings, floor plans. The system accepts as input digitized images of architectural plans and recognizes their spatial primitives (locations) and their spatial articulation on a variety of abstraction levels. The final output of chap is a description of the plan in terms of the grouping formations detected in its spatial articulation. The overall structure of the description is based on an analysis of its conformity to the formal rules of its “stylistic” context (which in the initial version of chap is classical architecture). Chapter 1 suggests that the poor performance of computerized architectural drawing and design systems is among others evidence of the necessity to computerize visual / spatial architectural representations. A recognition system such as chap offers comprehensive means for the investigation of a methodology for the development and use of such representations. Chapter 2 describes a fundamental task of chap: recognition of the position and shape of locations, the atomic parts of the description of an architectural plan in chap. This operation represents the final and most significant part of the first stage in processing an image input in machine environment. Chapter 3 moves to the next significant problem, recognition of the spatial arrangement of locations in an architectural plan, that is, recognition of grouping relationships that determine the subdivision of a plan into parts. In the absence of systematic and exhaustive typologic studies of classical architecture that would allow us to define a repertory of the location group types possible in classical architectural plans, Chapter 3 follows a bottom-up approach based on grouping relationships derived from elementary architectural knowledge and formalized with assistance from Gestalt theory and its antecedents. The grouping process described in Chapter 3 corresponds both in purpose and in structure to the derivation of a description of an image in computer vision [Marr 1982]. Chapter 4 investigates the well-formedness of the description of a classical architectural plan in an analytical manner: each relevant level (or sublevel) of the classical canon according to Tzonis & Lefaivre [1986] is transformed into a single group of criteria of well-formedness which is investigated independently. The hierarchical structure of the classical canon determines the coordination of these criteria into a sequence of cognitive filters which progressively analyses the correspondence of the descriptions derived as in Chapter 3 to the constraints of the canon. The methodology and techniques presented in the dissertation are primarily considered with respect to chap, a specific recognition system. The resulting specification of chap gives a measure of the use of such a system within the context of a computerized collection of architectural precedents and also presents several extensions to other areas of architecture. Although these extensions are not considered as verifiable claims, Chapter 5 describes some of their implications, including on the role of architectural drawing in computerized design systems, on architectural typologies, and on the nature and structure of generative systems in architecture.
series thesis:PhD
email
last changed 2003/02/12 22:37

_id e1d1
authors Shafer, Steven A. and Kanade, Takeo
year 1982
title Using Shadows in Finding Surface Orientations
source 61 p. : ill.` Pittsburgh, PA: Department of Computer Science, CMU, January, 1982. CMU-CS- 82-100
summary Given a line drawing from an image with shadow regions identified, the shapes of the shadows can be used to generate constraints on the orientations of the surfaces involved. This paper describes the theory which governs those constraints under orthography. A 'Basic Shadow Problem' is first posed, in which there is a single light source, and a single surface casts a shadow on another (background) surface. There are six parameters to determine: the orientation (2 parameters) for each surface, and the direction of the vector (2 parameters) pointing at the light source. If some set of 3 of these are given in advance, the remaining 3 can then be determined geometrically
keywords The solution method consists of identifying 'illumination surfaces' consisting of illumination vectors, assigning Huffman-Clowes line labels to
series CADline
last changed 2003/06/02 13:58

_id fafa
authors Marr, David
year 1982
title Vision: A Computational Investigation into the Human Representation and Processing of Visual Information
source New York, NY: W.H. Freeman and Company
summary Marr's demonstrations that retinal receptive field geometry could be derived by Fourier transformation of spatial frequency sensitivity data, that edges and contours could be detected by finding zero crossings in the light gradient by taking the Laplacian or second directional derivative, that excitatory and inhibitory receptive fields could be constructed from "DOG" functions (the difference of two Gaussians), and that the visual system used a two-dimensional convolution integral with a Gaussian prefilter as an operator for bandwidth optimation on the retinal light distribution, were more powerful than anything that had been seen up to that time. It was as if vision research suddenly acquired its own Principia Mathematica, or perhaps General Relativity Theory, in terms of the new explanatory power Marr's theories provided. Truly an extraordinary book from an extraordinary thinker in the area of perception, vision, and the brain.
series other
last changed 2003/04/23 15:14

_id e55f
authors Newell, Allen and Rosenbloom, Paul S.
year 1982
title Mechanisms of Skill Acquisition and the Law of Practice
source [2], III, 52 p. : ill Design Research Center, CMU, April, 1982. DRC-15-16-82. includes bibliography.
summary Practice, and the performance improvement that it engenders, has long been a major topic in psychology. In this paper, both experimental and theoretical approaches are employed in an investigation of the mechanisms underlying this improvement. On the experimental side, it is argued that a single law, the power of law of practice, adequately describes all of the practice data. On the theoretical side, a model of practice rooted in modern cognitive psychology, the chunking theory of learning, is formulated. The paper consists of (1) the presentation of a set of empirical practice curves; (2) mathematical investigations into the nature of power law functions; (3) evaluations of the ability of three different classes of functions to adequately model the empirical curves; (4) a discussion of the existing models of practice; (5) a presentation of the chunking theory of learning
keywords problem solving, learning, psychology, knowledge acquisition, practice, expert, representation
series CADline
last changed 2003/06/02 13:58

_id 2c13
authors Knuth, Donald E.
year 1982
title The Concept of a Meta-Font
source Visible Language. Winter, 1982. vol. XVI: pp. 3-27 : ill. includes bibliography
summary A single drawing of a single letter reveals only a small part of what was in the designer's mind when that letter was drawn. But when precise instructions are given about how to make such a drawing, the intelligence of that letter can be captured in a way that permits us to obtain an infinite variety of related letters from the same specification. Instead of merely describing a single letter, such instructions explain how that letter would change its shape if other parameters of the design were changed. Thus an entire font of letters and other symbols can be specified so that each character adapts itself to varying conditions in an appropriate way. Initial experiments with a precise language for pen motions suggest strongly that the font designer of the future should not simply design isolated alphabets; the challenge will be to explain exactly how each design should adapt itself gracefully to a wide range of changes in the specification. This paper gives examples of a meta-font and explains the changeable parameters in its design
keywords programming, computer graphics, parametrization
series CADline
last changed 2003/06/02 13:58

_id 8a88
authors Anderson, David P.
year 1982
title Hidden Line Elimination in Projected Grid Surfaces
source ACM Transactions on Graphics. October, 1982. vol. 1: pp. 274-288 : ill. includes a short bibliography
summary The hidden line and hidden surface problems are simpler when restricted to special classes of objects. An example is the class of grid surfaces, that is, graphs of bivariate functions represented by their values on a set of grid points. Projected grid surfaces have geometric properties which permit hidden line or hidden surface elimination to be done more easily than in the general case. These properties are discussed in this paper, and an algorithm is given which exploits them
keywords algorithms, hidden lines, hidden surfaces, grids, computer graphics
series CADline
last changed 2003/06/02 10:24

_id 2045
authors Balas, Egon
year 1982
title A Class of Location, Distribution and Scheduling Problems : Modeling and Solution Methods
source 21 p., 8 + 4 p. of appendix : ill. Carnegie Mellon University, Pittsburgh, PA: Design Research Center, December, 1982. includes bibliography
summary Discusses the potential of set covering techniques. Illustrates problem formulation techniques on several important classes of real-world problems. Also describes a class of algorithms for solving set covering problems, based on cutting planes, heuristics and subgradient optimization
keywords problem solving, methods, algorithms, problem definition, modeling, optimization, operations research
series CADline
last changed 1999/02/12 15:07

_id 898a
authors Bay, J.H.
year 2002
title Cognitive Biases and Precedent Knowledge in Human and Computer-Aided Design Thinking
source CAADRIA 2002 [Proceedings of the 7th International Conference on Computer Aided Architectural Design Research in Asia / ISBN 983-2473-42-X] Cyberjaya (Malaysia) 18–20 April 2002, pp. 213-220
doi https://doi.org/10.52842/conf.caadria.2002.213
summary Cognitive biases (illusions) and potential errors can occur when using precedent knowledge for analogical, pre-parametric and qualitative design thinking. This paper refers largely to part of a completed research (Bay 2001) on how heuristic biases, discussed by Tversky and Kahneman (1982) in cognitive psychology, can affect judgement and learning of facts from precedents in architectural design, made explicit using a kernel of conceptual system (Tzonis et. al., 1978) and a framework of architectural representation (Tzonis 1992). These are used here to consider how such illusions and errors may be transferred to computer aided design thinking.
series CAADRIA
email
last changed 2022/06/07 07:54

_id cf2003_m_040
id cf2003_m_040
authors BAY, Joo-Hwa
year 2003
title Making Rebuttals Available Digitally for Minimising Biases in Mental Judgements
source Digital Design - Research and Practice [Proceedings of the 10th International Conference on Computer Aided Architectural Design Futures / ISBN 1-4020-1210-1] Tainan (Taiwan) 13–15 October 2003, pp. 147-156
summary The problem of heuristic biases (illusions) discussed by Tversky and Kahneman (1982) that can lead to errors in judgement by human designers, when they use precedent knowledge presented graphically (Bay 2001). A Cognitive framework of belief, goal, and decision, and a framework of representation of architectural knowledge by Tzonis are used to map out the problem of heuristic biases in the human mind. These are used to discuss what aspects of knowledge can be presented explicitly and digitally to users to make rebuttal more available for human thinking at the cognitive level. The discussion is applicable to both inductive and analytic digital knowledge systems that use precedent knowledge. This discussion is targeted directly at means of addressing bias in the human mind using digital means. The problem of human bias in machine learning and generalisation are discussed in a different paper, and the problems of international or non-intentional machine bias are not part of discussion in this paper.
keywords analogy, bias, design thinking, environmental design, heuristics
series CAAD Futures
last changed 2003/11/22 07:26

_id 1b10
id 1b10
authors Bay, Joo-Hwa
year 2001
title Cognitive Biases - The case of tropical architecture
source Delft University of Technology
summary This dissertation investigates, i) How cognitive biases (or illusions) may lead to errors in design thinking, ii) Why architects use architectural precedents as heuristics despite such possible errors, and iii) Develops a design tool that can overcome this type of errors through the introduction of a rebuttal mechanism. The mechanism controls biases and improves accuracy in architectural thinking. // The research method applied is interdisciplinary. It employs knowledge from cognitive science, environmental engineering, and architectural theory. The case study approach is also used. The investigation is made in the case of tropical architecture. The investigation of architectural biases draws from work by A. Tversky and D. Kahneman in 1982 on “Heuristics and biases”. According to Tversky and Kahneman, the use of heuristics of representativeness (based on similarity) and availability (based on ease of recall and imaginability) for judgement of probability can result in cognitive biases of illusions of validity and biases due to imaginability respectively. This theory can be used analogically to understand how errors arise in the judgement of environmental behaviour anticipated from various spatial configurations, leading to designs with dysfunctional performances when built. Incomplete information, limited time, and human mental resources make design thinking in practice difficult and impossible to solve. It is not possible to analyse all possible alternative solutions, multiple contingencies, and multiple conflicting demands, as doing so will lead to combinatorial explosion. One of the ways to cope with the difficult design problem is to use precedents as heuristic devices, as shortcuts in design thinking, and at the risk of errors. This is done with analogical, pre-parametric, and qualitative means of thinking, without quantitative calculations. Heuristics can be efficient and reasonably effective, but may not always be good enough or even correct, because they can have associated cognitive biases that lead to errors. Several debiasing strategies are discussed, and one possibility is to introduce a rebuttal mechanism to refocus the designer’s thinking on the negative and opposite outcomes in his judgements, in order to debias these illusions. The research is carried out within the framework of design theory developed by the Design Knowledge System Research Centre, TUDelft. This strategy is tested with an experiment. The results show that the introduction of a rebuttal mechanism can debias and improve design judgements substantially in environmental control. The tool developed has possible applications in design practice and education, and in particular, in the designing of sustainable environments.
keywords Design bias; Design knowledge; Design rebuttal; Design Precedent; Pre-parametric design; Tropical architecture; Sustainability
series thesis:PhD
type normal paper
email
last changed 2006/05/28 07:42

_id 8239
authors Campello, Ruy Eduardo and Maculan, Nelson
year 1982
title On Deep Disjunctive Cutting Planes for Set Partitioning : A Computationally Oriented Research.
source Pittsburgh: Design Research Center, CMU [DRC-70-11-82], 10 p.
summary Several mathematical programming problems can be formulated as Disjunctive Programming Problems. This approach offers a powerful procedure for the generation of new and strong cutting planes with desirable properties. For general integer programs, the traditional cutting plane methodologies proved less efficient than enumerative techniques. However, for certain classes of problems, such as set partitioning, cutting planes are known to be efficient. Since the disjunctive cuts are strong, they can be expected to perform better. This paper reports on computational results with disjunctive B(.) cuts for the set partitioning problem, evaluated in terms of computer resources and other independent measures in solving specific randomly generated test problems under controlled conditions. [includes bibliography].
keywords Mathematics, Operations Research, Integer Programming, Optimization
series CADline
last changed 1999/02/15 15:17

_id 66df
authors Cendes, Z.J., Minhas, F.U. and Silvester, P.P.
year 1982
title Universal Finite Element Matrices for Tetrahedra
source 45, [22] p Pittsburgh: Design Research Center, CMU, December, 1982. DRC- 18-58-82. includes bibliography.
summary Methods are described for forming finite element matrices for a wide variety of operators on tetrahedral finite elements, in a manner similar to that previously employed for line segments and triangles. This technique models the differentiation and product-embedding operators as rectangular matrices, and produces finite element matrices by replacing all required analytic operations by their finite matrix analogues. The method is illustrated by deriving the conventional matrix representation for Laplace's equation. Brief computer programs are given, which generate universal finite element matrices for use in various applications
keywords mathematics, computational geometry, finite elements, analysis
series CADline
last changed 2003/06/02 13:58

_id 89e4
authors Cendes, Z.J., Shenton, D. and H. Shahnasser
year 1982
title Adaptive Finite Element Mesh Generation Using the Delaunay Algorithm
source 3 p. : ill. Pittsburgh: Design Research Center, CMU, December, 1982
summary Includes bibliography. A two-dimensional generator is described which automatically creates optimal finite element meshes using the Delaunay triangulation algorithm. The mesh generator is adaptive in the sense that elements containing the largest normalized errors are automatically refined, providing meshes with a uniform error density. The system runs on a PERQ computer made by Three Rivers Computer Company. It is menu oriented and utilizes multiple command and display windows to create and edit the object description interactively. Mesh generation from the object data base is automatic, although it may be modified interactively by the user if desired. Application of the mesh generator to electric machine design and to magnetic bubble simulation shows it to be one of the most powerful and easy to use systems yet devised
keywords electrical engineering, triangulation, algorithms, OOPS, finite elements, analysis
series CADline
last changed 2003/06/02 13:58

_id 666c
authors Clark, K. and Tarnlund, S. (eds.)
year 1982
title Logic Programming
source 366 p
summary Orlando, Florida: Academic Press,Inc., 1982. includes bibliography and index p. [341]-361. Based on the proceedings of the first international workshop on logic programming held in Debrecen, Hungry. ( A.P.I.C. Studies in Data Processing No.16). The 23 papers cover issues like application of logic programming, implementation, control, and understanding natural languages
keywords techniques, programming, logic, languages, symbolic, AI
series CADline
last changed 2002/12/13 15:23

_id adcd
authors Cohen, Paol R. and Feigenbaum, Edward A. (editors)
year 1982
title The Handbook of Artificial Intelligence
source xviii, 639 p. Stanford, California: HeurisTech Press, 1982. vol. 3: See The Handbook of Artificial intelligence edited by Avron, Barr and Feigenbaum, Edward. Includes bibliography p. 565-586 and cumulative indexes
summary Part of three volumes, this volume contains chapters on models of cognition, automatic deduction, vision, learning and planning
keywords AI, deduction, learning, cognition, planning
series CADline
last changed 2003/06/02 13:58

_id 482a
authors Cole, Sam
year 1982
title A Microprocessor Revolution and the World Distribution of Income: A General Equilibrium Approach
source International Political Science Review. 1982. vol.3: pp. 434- 454 ; ill. includes bibliography
summary This article shows that even if the world economy is able to withstand and surmount the present world crisis, the combination of market forces and rapid technical change that would be the result of a microprocessor revolution will give rise to large shifts in the distribution of income within and between both rich and poor countries. Some developed and developing economies may be unable to join the move to new technologies. In a world governed by only economic forces, all countries, whether they choose to adopt new systems of production or not, will be affected. Indeed, whatever their degree of involvement, all countries are beginning to feel in varying degrees the chain reaction that reverberates through and between all sectors of their domestic and the world economies. To gain insights into interrelations between technological change and global markets, this article uses a special type of model -- a general equilibrium model -- that enables the study to focus on exactly these variables
keywords technology, economics
series CADline
last changed 1999/02/12 15:07

_id sigradi2006_e183a
id sigradi2006_e183a
authors Costa Couceiro, Mauro
year 2006
title La Arquitectura como Extensión Fenotípica Humana - Un Acercamiento Basado en Análisis Computacionales [Architecture as human phenotypic extension – An approach based on computational explorations]
source SIGraDi 2006 - [Proceedings of the 10th Iberoamerican Congress of Digital Graphics] Santiago de Chile - Chile 21-23 November 2006, pp. 56-60
summary The study describes some of the aspects tackled within a current Ph.D. research where architectural applications of constructive, structural and organization processes existing in biological systems are considered. The present information processing capacity of computers and the specific software development have allowed creating a bridge between two holistic nature disciplines: architecture and biology. The crossover between those disciplines entails a methodological paradigm change towards a new one based on the dynamical aspects of forms and compositions. Recent studies about artificial-natural intelligence (Hawkins, 2004) and developmental-evolutionary biology (Maturana, 2004) have added fundamental knowledge about the role of the analogy in the creative process and the relationship between forms and functions. The dimensions and restrictions of the Evo-Devo concepts are analyzed, developed and tested by software that combines parametric geometries, L-systems (Lindenmayer, 1990), shape-grammars (Stiny and Gips, 1971) and evolutionary algorithms (Holland, 1975) as a way of testing new architectural solutions within computable environments. It is pondered Lamarck´s (1744-1829) and Weismann (1834-1914) theoretical approaches to evolution where can be found significant opposing views. Lamarck´s theory assumes that an individual effort towards a specific evolutionary goal can cause change to descendents. On the other hand, Weismann defended that the germ cells are not affected by anything the body learns or any ability it acquires during its life, and cannot pass this information on to the next generation; this is called the Weismann barrier. Lamarck’s widely rejected theory has recently found a new place in artificial and natural intelligence researches as a valid explanation to some aspects of the human knowledge evolution phenomena, that is, the deliberate change of paradigms in the intentional research of solutions. As well as the analogy between genetics and architecture (Estévez and Shu, 2000) is useful in order to understand and program emergent complexity phenomena (Hopfield, 1982) for architectural solutions, also the consideration of architecture as a product of a human extended phenotype can help us to understand better its cultural dimension.
keywords evolutionary computation; genetic architectures; artificial/natural intelligence
series SIGRADI
email
last changed 2016/03/10 09:49

For more results click below:

this is page 0show page 1show page 2show page 3HOMELOGIN (you are user _anon_501483 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002