CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers

Hits 1 to 20 of 75

_id 6f57
authors Searle, John R.
year 1980
title Minds, Brains, and Programs
source The Behavioral and Brain Sciences. Cambridge University Press., 1980. vol. 3: pp. 417-457. includes bibliography
summary This article can be viewed as an attempt to explore the consequences of two propositions: (1) Intentionallity in human beings (and animals) is a product of causal features of the brain. The author assumes this is an empirical fact about the actual causal relations between mental processes and brains. It says simply that certain brain processes are sufficient for intentionallity. (2) Instantiating a computer program is never by itself a sufficient condition of intentionallity. The main argument of this paper is directed at establishing this claim. The form of the argument is to show how a human agent could instantiate the program and still not have the relevant intentionallity. These two propositions have the following consequences: (3) The explanation of how the brain produces intentionallity cannot be that it does it by instantiating a computer program. This is a strict logical consequence of 1 and 2. (4) Any mechanism capable of producing intentionallity must have causal powers equal to those of the brain. This is meant to be a trivial consequence of 1. (5) Any attempt literally to create intentionallity artificially (strong AI) could not succeed just by designing programs but would have to duplicate the causal powers of the human brain. This follows from 2 and 4. 'Could a machine think?' On the argument advanced here only a machine could think, and only very special kinds of machines, namely brains and machines withÔ h) 0*0*0*°° ÔŒ internal causal powers equivalent to those of brains
keywords And that is why strong AI has little to tell us about thinking, since
series CADline
last changed 2003/06/02 08:24

_id 48db
authors Proctor, George
year 2001
title CADD Curriculum - The Issue of Visual Acuity
source Architectural Information Management [19th eCAADe Conference Proceedings / ISBN 0-9523687-8-1] Helsinki (Finland) 29-31 August 2001, pp. 192-200
summary Design educators attempt to train the eyes and minds of students to see and comprehend the world around them with the intention of preparing those students to become good designers, critical thinkers and ultimately responsible architects. Over the last eight years we have been developing the digital media curriculum of our architecture program with these fundamental values. We have built digital media use and instruction on the foundation of our program which has historically been based in physical model making. Digital modeling has gradually replaced the capacity of physical models as an analytical and thinking tool, and as a communication and presentation device. The first year of our program provides a foundation and introduction to 2d and 3d design and composition, the second year explores larger buildings and history, the third year explores building systems and structure through design studies of public buildings, fourth year explores urbanism, theory and technology through topic studios and, during the fifth year students complete a capstone project. Digital media and CADD have and are being synchronized with the existing NAAB accredited regimen while also allowing for alternative career options for students. Given our location in the Los Angeles region, many students with a strong background in digital media have gone on to jobs in video game design and the movie industry. Clearly there is much a student of architecture must learn to attain a level of professional competency. A capacity to think visually is one of those skills and is arguably a skill that distinguishes members of the visual arts (including Architecture) from other disciplines. From a web search of information posted by the American Academy of Opthamology, Visual Acuity is defined as an ability to discriminate fine details when looking at something and is often measured with the Snellen Eye Chart (the 20/20 eye test). In the context of this paper visual acuity refers to a subject’s capacity to discriminate useful abstractions in a visual field for the purposes of Visual Thinking- problem solving through seeing (Arnheim, 1969, Laseau 1980, Hoffman 1998). The growing use of digital media and the expanding ability to assemble design ideas and images through point-and-click methods makes the cultivation and development of visual skills all the more important to today’s crop of young architects. The advent of digital media also brings into question the traditional, static 2d methods used to build visual skills in a design education instead of promoting active 3d methods for teaching, learning and developing visual skills. Interactive digital movies provide an excellent platform for promoting visual acuity, and correlating the innate mechanisms of visual perception with the abstractions and notational systems used in professional discourse. In the context of this paper, pedagogy for building visual acuity is being considered with regard to perception of the real world, for example the visual survey of an environment, a site or a street scene and how that visual survey works in conjunction with practice.
keywords Curriculum, Seeing, Abstracting, Notation
series eCAADe
last changed 2001/08/06 20:38

_id 7f71
authors De Landa, Manuel
year 1998
title Meshworks, Hierarchies and Interfaces
source Virtual Dimension Ed. Beckmann, J., Princeton Architectural Press, NewYork
summary The world of interface design is today undergoing dramatic changes which in their impact promise to rival those brought about by the use of the point-and-click graphical interfaces popularized by the Macintosh in the early 1980's. The new concepts and metaphors which are aiming to replace the familiar desk-top metaphor all revolve around the notion of semi-autonomous, semi-intelligent software agents. To be sure, different researchers and commercial companies have divergent conceptions of what these agents should be capable of, and how they should interact with computer users. But whether one aims to give these software creatures the ability to learn about the users habits, as in the non-commercial research performed at MIT autonomous agents group, or to endow them with the ability to perform transactions in the users name, as in the commercial products pioneered by General Magic, the basic thrust seems to be in the direction of giving software programs more autonomy in their decision-making capabilities.
series other
last changed 2003/04/23 13:50

_id 6a59
authors Franklin, Randolph
year 1980
title A Linear Time Exact Hidden Surface Algorithm
source SIGGRAPH '80 Conference Proceedings. July, 1980. vol. 14 ; no. 3: pp. 117-133 : ill. includes bibliography
summary This Paper presents a new hidden surface algorithm. Its output is the set of the visible pieces of edges and faces, and is as accurate as the arithmetic precision of the computer. Thus calculating the hidden surfaces for a higher resolution device takes no more time. If the faces are independently and identically distributed, then the execution time is linear in the number of faces. In particular, the execution time does not increase with the depth complexity. This algorithm overlays a grid on the screen whose fineness depends on the number and size of the faces. Edges and faces are sorted into grid cells. Only objects in the same cell can intersect or hide each other. Also, if a face completely covers a cell then nothing behind it in the cell is relevant. Three programs have tested this algorithm. The first verified the variable grid concept on 50,000 intersecting edges. The second verified the linear time, fast speed, and irrelevance of depth complexity for hidden lines on 10,000 spheres. This also tested depth complexities up to 30, and showed that perspective scenes with the farther objects smaller are even faster to calculate. The third verified this for hidden surfaces on 3,000 squares
keywords hidden surfaces, algorithms, hidden lines, variables, grids, computer graphics, programming
series CADline
last changed 2003/06/02 11:58

_id b190
authors Goldberg, Adele and Robson, David
year 1983
title Smalltalk-80: The language and its implementation
source New York, NY: Addison Wesley Co
summary Smalltalk-80 is the classic standard Smalltalk language as described in Smalltalk-80: The Language and Its Implementation by Goldberg and Robson. This book is commonly called "the Blue Book". Squeak implements the dialect of Smalltalk described in this book, but has a different implementation. Overview of the Smalltalk Language Smalltalk is a general purpose, high level programming language. It was the first original "pure" object oriented language, but not the first to use the object oriented concept, which is credited to Simula 67. The explosive growth of Object Oriented Programming (OOP) technologies began in the early 1980's, with Smalltalk's introduction. Behind it was the idea that the individual human user should be the most important component of any computing system, and that programming should be a natural extension of thinking, and also a dynamic and evolutionary process consistent with the model of human learning activity. In Smalltalk, these ideas are embodied in a framework for human-computer communication. In a sense, Smalltalk is yet another language like C and Pascal, and programs can be written in Smalltalk that have the look and feel of such conventional languages. The difference lies * in the amount of code that can be reduced, * less cryptic syntax, * and code that is easier to handle for application maintenance and enhancement. But Smalltalk's most powerful feature is easy code reuse. Smalltalk makes reuse of programs, routines, and subroutines (methods) far easier. Though procedural languages allow reuse too, it is harder to do, and much easier to cheat. It is no surprise that Smalltalk is relatively easy to learn, mainly due to its simple syntax and semantics, as well as few concepts. Objects, classes, messages, and methods form the basis of programming in Smalltalk. The general methodology to use Smalltalk The notion of human-computer interface also results in Smalltalk promoting the development of safer systems. Errors in Smalltalk may be viewed as objects telling users that confusion exists as to how to perform a desired function.
series other
last changed 2003/04/23 13:14

_id 0439
authors Kant, Elaine
year 1980
title A Knowledge-Based Approach to Using Efficiency Estimation in Program Synthesis
source 1980? pp. 457-462. includes bibliography
summary This paper describes a system for using efficiency knowledge in program synthesis. The system, called LIBRA, uses a combination of knowledge-based rules and algebraic cost estimates to compare potential program implementations. Efficiency knowledge is used to control the selection of algorithm and data structure implementations and the application of optimizing transformations. Prototypes of programming constructs and of cost estimation techniques are used to simplify the efficiency analysis process and to assist in the acquisition of efficiency knowledge associated with new coding knowledge. LIBRA has been used to guide the selection of implementations for several programs that classify, retrieve information, sort, and generate prime numbers
keywords knowledge base, systems, programming, performance, synthesis, evaluation
series CADline
last changed 1999/02/12 14:08

_id a9d9
authors Rosenthal, David S.H.
year 1980
title Tools for Constructing User Interfaces
source 1980? 11 p. includes bibliography
summary A review of the tools available for easing the tasks of constructing and modifying the user interface of CAD application programs
keywords user interface, programming, tools, CAD
series CADline
last changed 2003/06/02 11:58

_id 9fcb
authors Steele, Guy Lewis
year 1980
title The Definition and Implementation of a Computer Programming Language Based Constraints
source MIT - AITR-595
summary The constraint paradigm is a model of computation in which values are deduced whenever possible, under the limitation that deductions be local in a certain sense. One may visualize a constraint 'program' as a network of devices connected by wires. Data values may flow along the wires, and computation is performed by the devices. A device computes using only locally available information (with a few exceptions), and places newly derived values on other, locally attached wires. In this way computed values are propagated. An advantage of the constraint paradigm (not unique to it) is that a single relationship can be used in more than one direction. The connections to a device are not labelled as inputs and outputs; a device will compute with whatever values are available, and produce as many new values as it can. General theorem provers are capable of such behavior, but tend to suffer from combinatorial explosion; it is not usually useful to derive all the possible consequences of a set of hypotheses. The constraint paradigm places a certain kind of limitation on the deduction process. The limitations imposed by the constraint paradigm are not the only one possible. It is argued, however, that they are restrictive enough to forestall combinatorial explosion in many interesting computational situations, yet permissive enough to allow useful computations in practical situations. Moreover, the paradigm is intuitive: It is easy to visualize the computational effects of these particular limitations, and the paradigm is a natural way of expressing programs for certain applications, in particular relationships arising in computer-aided design. A number of implementations of constraint-based programming languages are presented. A progression of ever more powerful languages is described, complete implementations are presented and design difficulties and alternatives are discussed. The goal approached, though not quite reached, is a complete programming system which will implicitly support the constraint paradigm to the same extent that LISP, say, supports automatic storage management.
series thesis:PhD
email gls@labean.East.Sun.COM
last changed 2003/02/12 21:37

_id fc80
authors Ubbelohde, S. and Humann, C.
year 1998
title Comparative Evaluation of Four Daylighting Software Programs
source 1998 ACEEE Summer Study on Energy Efficiency in Buildings Proceedings. American Council for an Energy-Efficient Economy
summary By the mid-1980's, a number of software packages were under development to predict daylighting performance in buildings, in particular illumination levels in daylighted spaces. An evaluation in 1988 by Ubbelohde et al. demonstrated that none of the software then available was capable of predicting the simplest of real daylighting designs. In the last ten years computer capabilities have evolved rapidly and we have four major packages widely available in the United States. This paper presents a comparative evaluation from the perspective of building and daylighting design practice. A contemporary building completed in 1993 was used as a base case for evaluation. We present the results from field measurements, software predictions and physical modeling as a basis for discussing the capabilities of the software packages in architectural design practice. We found the current software packages far more powerful and nuanced in their ability to predict daylight than previously. Some can accurately predict quantitative daylight performance under varying sky conditions and produce handsome and accurate visualizations of the space. The programs differ significantly, however, in their ease of use, modeling basis and the emphasis between quantitative predictions and visualization in the output.
series other
last changed 2003/04/23 13:50

_id 0830
authors Ball, A. A.
year 1980
title How to Make the Bicubic Patch Work Using Reparametrisation
source 1980 ? 11 p. includes bibliography
summary This paper comprises a series of examples in numerical surface definition, loosely strung together, to show the practical limitations of the bicubic patch and how they can be overcome by reparametrisation. The concept of reparametrisation is more general than that used in computer- aided geometric design insofar as the reparametrisation is modeled in addition to the basic parametric equation
keywords CAD, computational geometry, curved surfaces, parametrization
series CADline
last changed 2003/06/02 11:58

_id cf2011_p170
id cf2011_p170
authors Barros, Mário; Duarte José, Chaparro Bruno
year 2011
title Thonet Chairs Design Grammar: a Step Towards the Mass Customization of Furniture
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 181-200.
summary The paper presents the first phase of research currently under development that is focused on encoding Thonet design style into a generative design system using a shape grammar. The ultimate goal of the work is the design and production of customizable chairs using computer assisted tools, establishing a feasible practical model of the paradigm of mass customization (Davis, 1987). The current research step encompasses the following three steps: (1) codification of the rules describing Thonet design style into a shape grammar; (2) implementing the grammar into a computer tool as parametric design; and (3) rapid prototyping of customized chair designs within the style. Future phases will address the transformation of the Thonet’s grammar to create a new style and the production of real chair designs in this style using computer aided manufacturing. Beginning in the 1830’s, Austrian furniture designer Michael Thonet began experimenting with forming steam beech, in order to produce lighter furniture using fewer components, when compared with the standards of the time. Using the same construction principles and standardized elements, Thonet produced different chairs designs with a strong formal resemblance, creating his own design language. The kit assembly principle, the reduced number of elements, industrial efficiency, and the modular approach to furniture design as a system of interchangeable elements that may be used to assemble different objects enable him to become a pioneer of mass production (Noblet, 1993). The most paradigmatic example of the described vision of furniture design is the chair No. 14 produced in 1858, composed of six structural elements. Due to its simplicity, lightness, ability to be stored in flat and cubic packaging for individual of collective transportation, respectively, No. 14 became one of the most sold chairs worldwide, and it is still in production nowadays. Iconic examples of mass production are formally studied to provide insights to mass customization studies. The study of the shape grammar for the generation of Thonet chairs aimed to ensure rules that would make possible the reproduction of the selected corpus, as well as allow for the generation of new chairs within the developed grammar. Due to the wide variety of Thonet chairs, six chairs were randomly chosen to infer the grammar and then this was fine tuned by checking whether it could account for the generation of other designs not in the original corpus. Shape grammars (Stiny and Gips, 1972) have been used with sucesss both in the analysis as in the synthesis of designs at different scales, from product design to building and urban design. In particular, the use of shape grammars has been efficient in the characterization of objects’ styles and in the generation of new designs within the analyzed style, and it makes design rules amenable to computers implementation (Duarte, 2005). The literature includes one other example of a grammar for chair design by Knight (1980). In the second step of the current research phase, the outlined shape grammar was implemented into a computer program, to assist the designer in conceiving and producing customized chairs using a digital design process. This implementation was developed in Catia by converting the grammar into an equivalent parametric design model. In the third phase, physical models of existing and new chair designs were produced using rapid prototyping. The paper describes the grammar, its computer implementation as a parametric model, and the rapid prototyping of physical models. The generative potential of the proposed digital process is discussed in the context of enabling the mass customization of furniture. The role of the furniture designer in the new paradigm and ideas for further work also are discussed.
keywords Thonet; furniture design; chair; digital design process; parametric design; shape grammar
series CAAD Futures
last changed 2012/02/11 18:21

_id 2fdd
authors Barsky, Brian A. and Thomas, Spencer W.
year 1980
title Transpline Curve Representation System
source April, 1980. 19 p. : ill. includes bibliography
summary An interactive curve representation system has been developed based on the concept of transforming among several parametric spline curve formulations. The available formulations are the interpolatory spline, uniform B-spline, spline under tension, and NU-spline. The system implementation is described in the context of a sample design session
keywords computational geometry, curves, representation, splines
series CADline
last changed 2003/06/02 11:58

_id 8629
authors Barzilay, Amos
year 1980
title Human Problem Solving on Master Mind
source Carnegie Mellon University
summary The purpose of this work is to analyze the task of playing Master Mind and to examine subjects behaviors on solving that task. The methods and the ideas that are used in the work are the same found in the references for other tasks. The author wants to show that those ideas and methods can be used for that specific task as well. In other words, subjects behave in such a domain as an information processing system. [includes bibliography]
keywords Psychology, Problem Solving
series CADline
last changed 1999/02/15 14:10

_id e825
authors Baybars, Ilker and Eastman, Charles M.
year 1980
title Enumerating Architectural Arrangements by Generating Their Underlying Graphs
source Environment and Planning B. 1980. vol. 7: pp. 289- 310 : ill. includes bibliography. -- See also 'Enumerating Architectural Arrangements: Comment on a Recent Paper by Baybars and Eastman' by C.F. Earl
summary One mathematical correspondence to the partitioning of the plane is a Weighted Plane Graph (WPG). This paper first focuses on the systematic generation of WPGs, in a fashion similar to crystal growth. During this process, the WPGs are represented by adjacency matrices. The authors, thus, present a method for embedding the WPG in the plane, given its adjacency matrix. These graphs can, then, be mapped into floor plans. The common practice here is the use of the `geometric dual' of a WPG. The authors propose, instead, the use of the `Pseudogeometric dual' of a WPG directly to translate (part of) a design brief into alternative spatial layouts. Also discussed is the ability to create courtyards and/or circulation spaces given a specific WPG, without increasing the size of the problem
keywords enumeration, architecture, floor plans, graphs, design process, automation, algorithms, space allocation, CAD
series CADline
last changed 2003/05/17 08:15

_id cf2011_p127
id cf2011_p127
authors Benros, Deborah; Granadeiro Vasco, Duarte Jose, Knight Terry
year 2011
title Integrated Design and Building System for the Provision of Customized Housing: the Case of Post-Earthquake Haiti
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 247-264.
summary The paper proposes integrated design and building systems for the provision of sustainable customized housing. It advances previous work by applying a methodology to generate these systems from vernacular precedents. The methodology is based on the use of shape grammars to derive and encode a contemporary system from the precedents. The combined set of rules can be applied to generate housing solutions tailored to specific user and site contexts. The provision of housing to shelter the population affected by the 2010 Haiti earthquake illustrates the application of the methodology. A computer implementation is currently under development in C# using the BIM platform provided by Revit. The world experiences a sharp increase in population and a strong urbanization process. These phenomena call for the development of effective means to solve the resulting housing deficit. The response of the informal sector to the problem, which relies mainly on handcrafted processes, has resulted in an increase of urban slums in many of the big cities, which lack sanitary and spatial conditions. The formal sector has produced monotonous environments based on the idea of mass production that one size fits all, which fails to meet individual and cultural needs. We propose an alternative approach in which mass customization is used to produce planed environments that possess qualities found in historical settlements. Mass customization, a new paradigm emerging due to the technological developments of the last decades, combines the economy of scale of mass production and the aesthetics and functional qualities of customization. Mass customization of housing is defined as the provision of houses that respond to the context in which they are built. The conceptual model for the mass customization of housing used departs from the idea of a housing type, which is the combined result of three systems (Habraken, 1988) -- spatial, building system, and stylistic -- and it includes a design system, a production system, and a computer system (Duarte, 2001). In previous work, this conceptual model was tested by developing a computer system for existing design and building systems (Benr__s and Duarte, 2009). The current work advances it by developing new and original design, building, and computer systems for a particular context. The urgent need to build fast in the aftermath of catastrophes quite often overrides any cultural concerns. As a result, the shelters provided in such circumstances are indistinct and impersonal. However, taking individual and cultural aspects into account might lead to a better identification of the population with their new environment, thereby minimizing the rupture caused in their lives. As the methodology to develop new housing systems is based on the idea of architectural precedents, choosing existing vernacular housing as a precedent permits the incorporation of cultural aspects and facilitates an identification of people with the new housing. In the Haiti case study, we chose as a precedent a housetype called “gingerbread houses”, which includes a wide range of houses from wealthy to very humble ones. Although the proposed design system was inspired by these houses, it was decided to adopt a contemporary take. The methodology to devise the new type was based on two ideas: precedents and transformations in design. In architecture, the use of precedents provides designers with typical solutions for particular problems and it constitutes a departing point for a new design. In our case, the precedent is an existing housetype. It has been shown (Duarte, 2001) that a particular housetype can be encoded by a shape grammar (Stiny, 1980) forming a design system. Studies in shape grammars have shown that the evolution of one style into another can be described as the transformation of one shape grammar into another (Knight, 1994). The used methodology departs takes off from these ideas and it comprises the following steps (Duarte, 2008): (1) Selection of precedents, (2) Derivation of an archetype; (3) Listing of rules; (4) Derivation of designs; (5) Cataloguing of solutions; (6) Derivation of tailored solution.
keywords Mass customization, Housing, Building system, Sustainable construction, Life cycle energy consumption, Shape grammar
series CAAD Futures
last changed 2012/02/11 18:21

_id 8a27
authors Bentley, Jon L. and Carruthers, Wendy
year 1980
title Algorithms for Testing the Inclusion of Points in Polygons
source Allertorn Conference on Communication, Control and Computing (18th : 1980). (10) p. includes bibliography
summary Determining whether a given point lies inside or outside a simple polygon is an important problem in many applications, including computer vision systems and computer-assisted political redistricting systems. In this paper the authors give algorithms for inclusion problems that are efficient for polygons that are 'close to convex' in a certain precise sense. An empirical study of polygons that arise in several applications shows that typical polygons are indeed 'close to convex,' and a program implementing the algorithm shows that is extremely efficient on point sets of practical sizes
keywords point inclusion, polygons, algorithms, computational geometry
series CADline
last changed 2003/06/02 11:58

_id 4580
authors Borgerson, B. R. and Johnson, Robert H.
year 1980
title Beyond CAD to Computer Aided Engineering
source (8) p. : ill. Manufacturing Data Systems Incorporated, 1980? includes bibliography
summary Current CAD systems significantly aid the drafting function and many provide some aid to selected design activities. For the development of mechanical systems, much more can be done. Future systems will aid the interactive engineering process of design, analysis, control, documentation, and manufacturing engineering. Computer based systems which address this broader spectrum of engineering activities are referred to as `Computer Aided Engineering,' or `CAE,' systems. CAE systems will use volumetric techniques to create and evaluate the individual components of a machine design in conjunction with data base management schemas to support the interrelationships of the components of machines. This paper focuses on computer assistance to the engineering of mechanical systems
keywords mechanical engineering, CAE, solid modeling, objects
series CADline
last changed 2003/06/02 11:58

_id 0105
authors Bossan, Mario and Ronchi, Alfredo M.
year 1989
title Presentazione Esperienza Didattica del Dipartimento di Ingegneria dei Sistemi Edilizi e Territoriali - Politecnico di Milano
source CAAD: Education - Research and Practice [eCAADe Conference Proceedings / ISBN 87-982875-2-4] Aarhus (Denmark) 21-23 September 1989, pp. 9.8.1-9.8.19
summary Didactic and research experience developed at the "Dipartimento di Ingegneria dei Sistemi Edilizi e Territoriali del Politecnico di Milano" in the environment of Computer Aided Architectural Design (CAAD). From the early part of the 1980's, using initially at an experimental level the resources available at the departmental centre of calculation various applications of CAD techniques in the building sector have been effected at DISET (Dipartimento di Ingegneria del Politecnico di Milano). During 1983, after a three year period of experimenting with these systems, it was decided to organise and activate a small computer aided design centre, within the department, the use of which was reserved for dissertation and research students.

series eCAADe
last changed 2003/11/21 14:16

_id 8593
authors Braid, I.C.
year 1980
title Superficial Blends in Geometric Modelling
source 12 p. : ill. February, 1980. Document No. 105. includes bibliography
summary In engineering practice, many blended surfaces are both indicated and manufactured by rounding off a sharp edge. The authors term them `superficial' blends in contrast to `designed' blends for which existing surface techniques are appropriate. The provision of superficial blends in a geometric modeling system is explained, and a method is given for drawing objects containing blended edges
keywords geometric modeling, CAD, representation, solids
series CADline
last changed 2003/06/02 11:58

_id 0189
authors Brodlie, K.W. (editor)
year 1980
title Mathematical Methods in Computer Graphics and Design
source xi, 147 p. : ill. New York: Academic Press, 1980. includes subject index
summary Based on the proceeding of the conference on mathematical methods in computer graphics and design, organized by the Institute of Mathematics and Its Applications and held at the university of Leicester on september 28th, 1978
keywords algorithms, geometric modeling, techniques, computer graphics, mathematics
series CADline
last changed 2003/06/02 11:58

For more results click below:

this is page 0show page 1show page 2show page 3HOMELOGIN (you are user _anon_828991 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002