CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 79

_id e952
authors Carrara, Gianfranco and Paoluzzi, Alberto
year 1980
title A Systems Approach to Building Program Planning
source computer Aided Building Design Laboratory Research Report. 80 p. : ill. Rome, Italy: December, 1980. CABD LAB RR. 80-02. includes bibliography
summary In this paper problems of design performance and of building program planning are considered from the view point of the general system theory. After having formalized the concept of requirement, performance and performance specification, it is shown that a set of building objects (spaces and constructive elements) foreseeable within a program is a semilattice, and that therefore the ordering of constructive elements and spaces corresponds to an ordering of relations among feasible 'behaviors.' A set of feasible behaviors is then presented as an abstract system, eventually discussing some assumptions on which to base an input-state-output representation of it
keywords theory, methods, problem solving, architecture, design, knowledge
series CADline
last changed 2003/06/02 13:58

_id a587
authors Cohen, Elaine, Lyche, Tom and Riesenfeld, Richard F.
year 1980
title Discrete B-Splines and Subdivision Techniques in Computer-Aided Geometric Design and Computer Graphics
source computer Graphics and Image Processing. October, 1980. vol. 14: pp. 87- 111 : ill. includes bibliography
summary The relevant theory of discrete B-splines with associated new algorithms is extended to provide a framework for understanding and implementing general subdivision schemes for nonuniform B-splines. The new derived polygon corresponding to an arbitrary refinement of the knot vector for an existing B-spline curve, including multiplicities, is shown to be formed by successive evaluations of the discrete B-spline defined by the original vertices, the original knot vector, and the refined knot vector. Existing subdivision algorithms can be seen as proper special cases. General subdivision has widespread applications in computer-aided geometric design, computer graphics, and numerical analysis. The new algorithms resulting from the new theory lead to a unification of the display model, the analysis model, and other needed models into a single geometric model from which other necessary models are easily derived. New sample algorithms for interference calculation, contouring, surface rendering, and other important calculations are presented
keywords computational geometry, theory, algorithms, computer graphics, B-splines, curved surfaces
series CADline
last changed 2003/06/02 13:58

_id acadia07_164
id acadia07_164
authors Diniz, Nancy; Turner, Alasdair
year 2007
title Towards a Living Architecture
doi https://doi.org/10.52842/conf.acadia.2007.164
source Expanding Bodies: Art • Cities• Environment [Proceedings of the 27th Annual Conference of the Association for Computer Aided Design in Architecture / ISBN 978-0-9780978-6-8] Halifax (Nova Scotia) 1-7 October 2007, 164-173
summary Interaction is the latest currency in architecture, as responsive components are now reacting to the inhabitant of the space. These components are designed and installed by the architect with a view to the phenomenology of space, where the experience of the environment is previewed and pre-constructed before it is translated into the conception of the space. However, this traditional approach to new technology leaves no scope for the architecture to be alive in and of itself, and thus the installed piece quickly becomes just that—an installation: isolated and uncontained by its environment. In this paper, we argue that a way to approach a responsive architecture is to design for a piece that is truly living, and in order to propose a living architecture first we need to understand what the architecture of a living system is. This paper suggests a conceptual framework based on the theory of Autopoiesis in order to create a “self-producing” system through an experiment entitled, “The Life of a Wall” (Maturana and Varela 1980). The wall has a responsive membrane controlled by a genetic algorithm that reconfigures its behaviour and learns to adapt itself continually to the evolutionary properties of the environment, thus becoming a situated, living piece.
series ACADIA
email
last changed 2022/06/07 07:55

_id a5c3
authors Er, M.C.
year 1981
title The Relations of the Computation of Fibonnaci Numbers with the Polyphase Sort
source 8 p. Wollongong: Department of Computing Science, University of Wollongong, September, 1981. includes bibliography
summary The theory of polyphase sort has simplified the mathematical derivations of Wilson and Shortt's (1980) algorithm, and offered an intuitive explanation of why Gries and Levin's (1980), and Urbanek's (1980) algorithms work. The computation of order-k Fibonacci numbers is equivalent to moving a window of matrix upwards in a series of ideal distributions
keywords Fibonacci, sorting, mathematics, algorithms
series CADline
last changed 2003/06/02 13:58

_id 4990
authors Gips, J. and Stiny, G.
year 1980
title Production System and Grammars: a Uniform Characterization
source Environment and Planning B. 1980. vol. 7: pp. 399-408 : tables. includes bibliography
summary The common structure underlying production system formalisms is developed. A variety of production system formalisms are summarized in terms of this structure. The structure is useful both for understanding existing systems and for developing new ones
keywords systems, shape grammars, theory
series CADline
last changed 2003/06/02 13:58

_id b190
authors Goldberg, Adele and Robson, David
year 1983
title Smalltalk-80: The language and its implementation
source New York, NY: Addison Wesley Co
summary Smalltalk-80 is the classic standard Smalltalk language as described in Smalltalk-80: The Language and Its Implementation by Goldberg and Robson. This book is commonly called "the Blue Book". Squeak implements the dialect of Smalltalk described in this book, but has a different implementation. Overview of the Smalltalk Language Smalltalk is a general purpose, high level programming language. It was the first original "pure" object oriented language, but not the first to use the object oriented concept, which is credited to Simula 67. The explosive growth of Object Oriented Programming (OOP) technologies began in the early 1980's, with Smalltalk's introduction. Behind it was the idea that the individual human user should be the most important component of any computing system, and that programming should be a natural extension of thinking, and also a dynamic and evolutionary process consistent with the model of human learning activity. In Smalltalk, these ideas are embodied in a framework for human-computer communication. In a sense, Smalltalk is yet another language like C and Pascal, and programs can be written in Smalltalk that have the look and feel of such conventional languages. The difference lies * in the amount of code that can be reduced, * less cryptic syntax, * and code that is easier to handle for application maintenance and enhancement. But Smalltalk's most powerful feature is easy code reuse. Smalltalk makes reuse of programs, routines, and subroutines (methods) far easier. Though procedural languages allow reuse too, it is harder to do, and much easier to cheat. It is no surprise that Smalltalk is relatively easy to learn, mainly due to its simple syntax and semantics, as well as few concepts. Objects, classes, messages, and methods form the basis of programming in Smalltalk. The general methodology to use Smalltalk The notion of human-computer interface also results in Smalltalk promoting the development of safer systems. Errors in Smalltalk may be viewed as objects telling users that confusion exists as to how to perform a desired function.
series other
last changed 2003/04/23 15:14

_id 76ce
authors Grimson, W.
year 1985
title Computational Experiments with a Feature Based Stereo Algorithm
source IEEE Trans. Pattern Anal. Machine Intell., Vol. PAMI-7, No. 1
summary Computational models of the human stereo system' can provide insight into general information processing constraints that apply to any stereo system, either artificial or biological. In 1977, Marr and Poggio proposed one such computational model, that was characterized as matching certain feature points in difference-of-Gaussian filtered images, and using the information obtained by matching coarser resolution representations to restrict the search'space for matching finer resolution representations. An implementation of the algorithm and'its testing on a range of images was reported in 1980. Since then a number of psychophysical experiments have suggested possible refinements to the model and modifications to the algorithm. As well, recent computational experiments applying the algorithm to a variety of natural images, especially aerial photographs, have led to a number of modifications. In this article, we present a version of the Marr-Poggio-Gfimson algorithm that embodies these modifications and illustrate its performance on a series of natural images.
series journal paper
last changed 2003/04/23 15:14

_id ddss2006-hb-187
id DDSS2006-HB-187
authors Lidia Diappi and Paola Bolchi
year 2006
title Gentrification Waves in the Inner-City of Milan - A multi agent / cellular automata model based on Smith's Rent Gap theory
source Van Leeuwen, J.P. and H.J.P. Timmermans (eds.) 2006, Innovations in Design & Decision Support Systems in Architecture and Urban Planning, Dordrecht: Springer, ISBN-10: 1-4020-5059-3, ISBN-13: 978-1-4020-5059-6, p. 187-201
summary The aim of this paper is to investigate the gentrification process by applying an urban spatial model of gentrification, based on Smith's (1979; 1987; 1996) Rent Gap theory. The rich sociological literature on the topic mainly assumes gentrification to be a cultural phenomenon, namely the result of a demand pressure of the suburban middle and upper class, willing to return to the city (Ley, 1980; Lipton, 1977, May, 1996). Little attempt has been made to investigate and build a sound economic explanation on the causes of the process. The Rent Gap theory (RGT) of Neil Smith still represents an important contribution in this direction. At the heart of Smith's argument there is the assumption that gentrification takes place because capitals return to the inner city, creating opportunities for residential relocation and profit. This paper illustrates a dynamic model of Smith's theory through a multi-agent/ cellular automata system approach (Batty, 2005) developed on a Netlogo platform. A set of behavioural rules for each agent involved (homeowner, landlord, tenant and developer, and the passive 'dwelling' agent with their rent and level of decay) are formalised. The simulations show the surge of neighbouring degradation or renovation and population turn over, starting with different initial states of decay and estate rent values. Consistent with a Self Organized Criticality approach, the model shows that non linear interactions at local level may produce different configurations of the system at macro level. This paper represents a further development of a previous version of the model (Diappi, Bolchi, 2005). The model proposed here includes some more realistic factors inspired by the features of housing market dynamics in the city of Milan. It includes the shape of the potential rent according to city form and functions, the subdivision in areal submarkets according to the current rents, and their maintenance levels. The model has a more realistic visualisation of the city and its form, and is able to show the different dynamics of the emergent neighbourhoods in the last ten years in Milan.
keywords Multi agent systems, Housing market, Gentrification, Emergent systems
series DDSS
last changed 2006/08/29 12:55

_id 0565
authors Oxman, Robert and Oxman, Rivka
year 1990
title The Computability of Architectural Knowledge
source The Electronic Design Studio: Architectural Knowledge and Media in the Computer Era [CAAD Futures ‘89 Conference Proceedings / ISBN 0-262-13254-0] Cambridge (Massachusetts / USA), 1989, pp. 171-185
summary In an important contribution to the theoretical foundation of design computing, Mitchell noted "an increasingly urgent need to establish a demonstrably sound, comprehensive, rigorously formalized theoretical foundation upon which to base practical software development efforts" (Mitchell, 1986). In this paper we propose such a theoretical framework. A basic assumption of this work is that the advancement of design computing is dependent upon the emergence of a rigorous formulation of knowledge in design. We present a model of knowledge in architectural design which suggests a promising conceptual basis for dealing with knowledge in computer-aided design systems. We require models which can represent the formal knowledge and manipulative operations of the designer in all of their complexity-that is formal models rather than just geometric models. Shape Grammars (Stiny,1980) represent an example of such models, and constitute a relatively high level of design knowledge as compared to, for example, use of symmetry operations to generate simple formal configurations. Building upon an understanding of the classes of design knowledge as the conceptual basis for formal modeling systems may contribute a new realization of the potential of the medium for design. This will require a comprehensive approach to the definition of architectural and design knowledge. We consider here the implications of a well-defined body of architectural and design knowledge for design education and the potential mutual interaction-in a knowledge-rich environment-of design learning and CAAD learning. The computational factors connected with the representation of design knowledge and its integration in design systems are among the key problems of CAAD. Mitchell's model of knowledge in design incorporates formal knowledge in a comprehensive, multi-level, hierarchical structure in which types of knowledge are correlated with computational concepts. In the main focus of this paper we present a structured, multi-level model of design knowledge which we discuss with respect to current architectural theoretical considerations. Finally, we analyze the computational and educational relevance of such models.
series CAAD Futures
email
last changed 2003/05/16 20:58

_id 2006_252
id 2006_252
authors Penttilä, Hannu
year 2006
title Managing the Changes within the Architectural Practice - The Effects of Information and Communication Technology (ICT)
doi https://doi.org/10.52842/conf.ecaade.2006.252
source Communicating Space(s) [24th eCAADe Conference Proceedings / ISBN 0-9541183-5-9] Volos (Greece) 6-9 September 2006, pp. 252-260
summary The architectural working environment has changed during the last 30 years more than ever before. Most of the changes have been related with information and communication technologies (ICT). Architectural working methods and tools have changed profoundly, when CAD has replaced more traditional methods and tools. Communicative working environment and document management within design & construction has also been changed to digital, meaning email and project webs. Completing a traditional architectural profile of the 20th centrury, a drawer-designer, contemporary communicating and managing skills plus mastering ICT are needed today to operate modern architectural practise properly. The objective of this study is to create a change-oriented understanding of the contemporary architectural profession concentrating on architectural information management. The first phase, a literature study, will be followed by interviews and case-studies, to examine three hypothetically different periods of time: a) 1980-85 the era before CAD, the last days of hand-drawing, b) 1993-98 the era of digital drawing, the expansion of architectural CAD, c) 2000-05 the rise of integrated and pervasive web-supported digital design. The study will propose new aspects to be included in the modern architectural profile, namingly project coordination, collaborative team-work, design information integration and profound digital content management.
keywords architectural profession; design practice; architectural ICT; change management
series eCAADe
email
last changed 2022/06/07 08:00

_id 48db
authors Proctor, George
year 2001
title CADD Curriculum - The Issue of Visual Acuity
doi https://doi.org/10.52842/conf.ecaade.2001.192
source Architectural Information Management [19th eCAADe Conference Proceedings / ISBN 0-9523687-8-1] Helsinki (Finland) 29-31 August 2001, pp. 192-200
summary Design educators attempt to train the eyes and minds of students to see and comprehend the world around them with the intention of preparing those students to become good designers, critical thinkers and ultimately responsible architects. Over the last eight years we have been developing the digital media curriculum of our architecture program with these fundamental values. We have built digital media use and instruction on the foundation of our program which has historically been based in physical model making. Digital modeling has gradually replaced the capacity of physical models as an analytical and thinking tool, and as a communication and presentation device. The first year of our program provides a foundation and introduction to 2d and 3d design and composition, the second year explores larger buildings and history, the third year explores building systems and structure through design studies of public buildings, fourth year explores urbanism, theory and technology through topic studios and, during the fifth year students complete a capstone project. Digital media and CADD have and are being synchronized with the existing NAAB accredited regimen while also allowing for alternative career options for students. Given our location in the Los Angeles region, many students with a strong background in digital media have gone on to jobs in video game design and the movie industry. Clearly there is much a student of architecture must learn to attain a level of professional competency. A capacity to think visually is one of those skills and is arguably a skill that distinguishes members of the visual arts (including Architecture) from other disciplines. From a web search of information posted by the American Academy of Opthamology, Visual Acuity is defined as an ability to discriminate fine details when looking at something and is often measured with the Snellen Eye Chart (the 20/20 eye test). In the context of this paper visual acuity refers to a subject’s capacity to discriminate useful abstractions in a visual field for the purposes of Visual Thinking- problem solving through seeing (Arnheim, 1969, Laseau 1980, Hoffman 1998). The growing use of digital media and the expanding ability to assemble design ideas and images through point-and-click methods makes the cultivation and development of visual skills all the more important to today’s crop of young architects. The advent of digital media also brings into question the traditional, static 2d methods used to build visual skills in a design education instead of promoting active 3d methods for teaching, learning and developing visual skills. Interactive digital movies provide an excellent platform for promoting visual acuity, and correlating the innate mechanisms of visual perception with the abstractions and notational systems used in professional discourse. In the context of this paper, pedagogy for building visual acuity is being considered with regard to perception of the real world, for example the visual survey of an environment, a site or a street scene and how that visual survey works in conjunction with practice.
keywords Curriculum, Seeing, Abstracting, Notation
series eCAADe
email
last changed 2022/06/07 08:00

_id 08ba
authors Requicha, Aristides A.G.
year 1980
title Representations for Rigid Solids : Theory, Methods, and Systems
source Computing Surveys December, 1980. vol. 12: pp. 437-464 : ill. includes bibliography.
summary Computer-based systems for modeling the geometry of rigid solid objects are becoming increasingly important in mechanical and civil engineering, architecture, computer graphics, computer vision, and other fields that deal with spatial phenomena. At the heart of such systems are symbol structures (representations) designating 'abstract solids' (subsets of Euclidean space) that model physical solids. Representations are the sources of data for procedures which compute useful properties of objects. The variety and uses of systems embodying representations of solids are growing rapidly, but so are the difficulties in assessing current designs, specifying the characteristics that future systems should exhibit, and designing systems to meet such specifications. This paper resolves many of these difficulties by providing a coherent view, based on sound theoretical principles, of what is presently known about the representation of solids. The paper is divided into three parts. The first introduces a simple mathematical framework for characterizing certain important aspects of representations, for example, their semantic (geometric) integrity. The second part uses the framework to describe and compare all of the major known schemes for representing solids. The third part briefly surveys extant geometric modeling systems and then applies the concepts developed in the paper to the high-level design of a multiple- representation geometric modeling system which exhibits a level of reliability and versatility superior to that of systems currently used in industrial computer-aided design and manufacturing
keywords CAD, CAM, computational geometry, geometric modeling, representation,CSG, B-rep
series CADline
last changed 2003/06/02 13:58

_id c380
authors Saaty, Thomas L. and Beltran, Miguel H.
year 1980
title Architectural Design by the Analytic Hierarchy Process
source Design Methods and Theories. 1980? vol. 14: pp. 124-134 : ill. and tables. include some bibliographical notes
summary The Analytic Hierarchy Process (AHP), which is explained and applied in another article in the same journal, is used to illustrate the architectural design of a house for a family of three. The plan of the house was develop to satisfy the family's needs by considering the size of the lot, the size and shape of the different architectural spaces, their priorities, and their overall contiguity. The salient feature of this paper is to show how closely the final plan, shape, size, geometric design and location relate to our mental criteria and personal needs, from the identification of the needs to the final plan. This approach also permits one to treat all these needs and their relation to the environment in a coherent framework
keywords architecture, design, methods
series CADline
last changed 2003/06/02 13:58

_id 952f
authors Soloway, E., Guzdial, M. and Hay, K.
year 1994
title Learner-Centered Design: The Challenge for HCI in the 21st Century
source Interactions , no. April (1994): 36-48
summary In the 1980's a major transformation took place in the computing world: attention was finally being paid to making computers easier-to-use. You know the history: in the 1970's folks at Xerox were exploring so-called personal computers and developing graphical, point-and-click interfaces. The goal was to make using computers less cognitively taxing, there- by permitting the user to focus more mental cycles on getting the job done. For some time people had recognized that there would be benefits if users could interact with computers using visual cues and motor movements instead of testu- al/linguistic strings. However, computer cycles were costly; they could hardly be wasted on supporting a non-textual interface. There was barely enough zorch (i.e., computer power, measured in your favorite unit) to simply calculate the payroll.
series journal paper
last changed 2003/04/23 15:50

_id c153
authors Ullman, Jeffrey D.
year 1980
title Principles of Database Systems
source 379 p. : ill. Potomac, Maryland: Computer Science Press, 1980. includes bibliography: p.360-371 and index. -- (Computer Software Engineering Series)
summary A large part is a description of relations, their algebra and calculus, and the query languages that have been designed using these concepts. There are explanations of how the theory can be used to design good systems. A description of the optimization of queries in relation-based query languages is provided, and a chapter is devoted to the recently developed protocols for guaranteeing consistency in databases that are operated on by many processes concurrently
keywords database, systems, theory, relational database, education, techniques
series CADline
last changed 2003/06/02 13:58

_id 40ad
authors Yessios, Chris I.
year 1980
title Generation and Visualization of Architectural Forms with Tekton
source 1980? pp. 68-79 : ill. includes bibliography
summary Tekton is an interactive computer aided architectural design software system. It incorporates graphic input and 3-D modeling capabilities, a potent notational system which is based on an algebra like linguistic model for the representation of transformation and spatial compositions, hidden face elimination, shadowing and texture rendering. The latter feature has been specifically designed for the visualization of architectural forms and materials, through renderings of a free hand drawing quality. They are derived by generative semi-random models, included in the system. The Tekton language allows for interactive unlimited editing and modification of previously generated compositions
keywords CAD, architecture, modeling, computer graphics, rendering
series CADline
last changed 2003/06/02 13:58

_id 0830
authors Ball, A. A.
year 1980
title How to Make the Bicubic Patch Work Using Reparametrisation
source 1980 ? 11 p. includes bibliography
summary This paper comprises a series of examples in numerical surface definition, loosely strung together, to show the practical limitations of the bicubic patch and how they can be overcome by reparametrisation. The concept of reparametrisation is more general than that used in computer- aided geometric design insofar as the reparametrisation is modeled in addition to the basic parametric equation
keywords CAD, computational geometry, curved surfaces, parametrization
series CADline
last changed 2003/06/02 13:58

_id cf2011_p170
id cf2011_p170
authors Barros, Mário; Duarte José, Chaparro Bruno
year 2011
title Thonet Chairs Design Grammar: a Step Towards the Mass Customization of Furniture
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 181-200.
summary The paper presents the first phase of research currently under development that is focused on encoding Thonet design style into a generative design system using a shape grammar. The ultimate goal of the work is the design and production of customizable chairs using computer assisted tools, establishing a feasible practical model of the paradigm of mass customization (Davis, 1987). The current research step encompasses the following three steps: (1) codification of the rules describing Thonet design style into a shape grammar; (2) implementing the grammar into a computer tool as parametric design; and (3) rapid prototyping of customized chair designs within the style. Future phases will address the transformation of the Thonet’s grammar to create a new style and the production of real chair designs in this style using computer aided manufacturing. Beginning in the 1830’s, Austrian furniture designer Michael Thonet began experimenting with forming steam beech, in order to produce lighter furniture using fewer components, when compared with the standards of the time. Using the same construction principles and standardized elements, Thonet produced different chairs designs with a strong formal resemblance, creating his own design language. The kit assembly principle, the reduced number of elements, industrial efficiency, and the modular approach to furniture design as a system of interchangeable elements that may be used to assemble different objects enable him to become a pioneer of mass production (Noblet, 1993). The most paradigmatic example of the described vision of furniture design is the chair No. 14 produced in 1858, composed of six structural elements. Due to its simplicity, lightness, ability to be stored in flat and cubic packaging for individual of collective transportation, respectively, No. 14 became one of the most sold chairs worldwide, and it is still in production nowadays. Iconic examples of mass production are formally studied to provide insights to mass customization studies. The study of the shape grammar for the generation of Thonet chairs aimed to ensure rules that would make possible the reproduction of the selected corpus, as well as allow for the generation of new chairs within the developed grammar. Due to the wide variety of Thonet chairs, six chairs were randomly chosen to infer the grammar and then this was fine tuned by checking whether it could account for the generation of other designs not in the original corpus. Shape grammars (Stiny and Gips, 1972) have been used with sucesss both in the analysis as in the synthesis of designs at different scales, from product design to building and urban design. In particular, the use of shape grammars has been efficient in the characterization of objects’ styles and in the generation of new designs within the analyzed style, and it makes design rules amenable to computers implementation (Duarte, 2005). The literature includes one other example of a grammar for chair design by Knight (1980). In the second step of the current research phase, the outlined shape grammar was implemented into a computer program, to assist the designer in conceiving and producing customized chairs using a digital design process. This implementation was developed in Catia by converting the grammar into an equivalent parametric design model. In the third phase, physical models of existing and new chair designs were produced using rapid prototyping. The paper describes the grammar, its computer implementation as a parametric model, and the rapid prototyping of physical models. The generative potential of the proposed digital process is discussed in the context of enabling the mass customization of furniture. The role of the furniture designer in the new paradigm and ideas for further work also are discussed.
keywords Thonet; furniture design; chair; digital design process; parametric design; shape grammar
series CAAD Futures
email
last changed 2012/02/11 19:21

_id 2fdd
authors Barsky, Brian A. and Thomas, Spencer W.
year 1980
title Transpline Curve Representation System
source April, 1980. 19 p. : ill. includes bibliography
summary An interactive curve representation system has been developed based on the concept of transforming among several parametric spline curve formulations. The available formulations are the interpolatory spline, uniform B-spline, spline under tension, and NU-spline. The system implementation is described in the context of a sample design session
keywords computational geometry, curves, representation, splines
series CADline
last changed 2003/06/02 13:58

_id 8629
authors Barzilay, Amos
year 1980
title Human Problem Solving on Master Mind
source Carnegie Mellon University
summary The purpose of this work is to analyze the task of playing Master Mind and to examine subjects behaviors on solving that task. The methods and the ideas that are used in the work are the same found in the references for other tasks. The author wants to show that those ideas and methods can be used for that specific task as well. In other words, subjects behave in such a domain as an information processing system. [includes bibliography]
keywords Psychology, Problem Solving
series CADline
last changed 1999/02/15 15:10

For more results click below:

this is page 0show page 1show page 2show page 3HOMELOGIN (you are user _anon_967326 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002