CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 82

_id b04c
authors Goerger, S., Darken, R., Boyd, M., Gagnon, T., Liles, S., Sullivan, J. and Lawson, J.
year 1996
title Spatial Knowledge Acquisition from Maps and Virtual Environments in Complex Architectural Space
source Proc. 16 th Applied Behavioral Sciences Symposium, 22-23 April, U.S. Airforce Academy, Colorado Springs, CO., 1996, 6-10
summary It has often been suggested that due to its inherent spatial nature, a virtual environment (VE) might be a powerful tool for spatial knowledge acquisition of a real environment, as opposed to the use of maps or some other two-dimensional, symbolic medium. While interesting from a psychological point of view, a study of the use of a VE in lieu of a map seems nonsensical from a practical point of view. Why would the use of a VE preclude the use of a map? The more interesting investigation would be of the value added of the VE when used with a map. If the VE could be shown to substantially improve navigation performance, then there might be a case for its use as a training tool. If not, then we have to assume that maps continue to be the best spatial knowledge acquisition tool available. An experiment was conducted at the Naval Postgraduate School to determine if the use of an interactive, three-dimensional virtual environment would enhance spatial knowledge acquisition of a complex architectural space when used in conjunction with floor plan diagrams. There has been significant interest in this research area of late. Witmer, Bailey, and Knerr (1995) showed that a VE was useful in acquiring route knowledge of a complex building. Route knowledge is defined as the procedural knowledge required to successfully traverse paths between distant locations (Golledge, 1991). Configurational (or survey) knowledge is the highest level of spatial knowledge and represents a map-like internal encoding of the environment (Thorndyke, 1980). The Witmer study could not confirm if configurational knowledge was being acquired. Also, no comparison was made to a map-only condition, which we felt is the most obvious alternative. Comparisons were made only to a real world condition and a symbolic condition where the route is presented verbally.
series other
last changed 2003/04/23 15:50

_id 08ba
authors Requicha, Aristides A.G.
year 1980
title Representations for Rigid Solids : Theory, Methods, and Systems
source Computing Surveys December, 1980. vol. 12: pp. 437-464 : ill. includes bibliography.
summary Computer-based systems for modeling the geometry of rigid solid objects are becoming increasingly important in mechanical and civil engineering, architecture, computer graphics, computer vision, and other fields that deal with spatial phenomena. At the heart of such systems are symbol structures (representations) designating 'abstract solids' (subsets of Euclidean space) that model physical solids. Representations are the sources of data for procedures which compute useful properties of objects. The variety and uses of systems embodying representations of solids are growing rapidly, but so are the difficulties in assessing current designs, specifying the characteristics that future systems should exhibit, and designing systems to meet such specifications. This paper resolves many of these difficulties by providing a coherent view, based on sound theoretical principles, of what is presently known about the representation of solids. The paper is divided into three parts. The first introduces a simple mathematical framework for characterizing certain important aspects of representations, for example, their semantic (geometric) integrity. The second part uses the framework to describe and compare all of the major known schemes for representing solids. The third part briefly surveys extant geometric modeling systems and then applies the concepts developed in the paper to the high-level design of a multiple- representation geometric modeling system which exhibits a level of reliability and versatility superior to that of systems currently used in industrial computer-aided design and manufacturing
keywords CAD, CAM, computational geometry, geometric modeling, representation,CSG, B-rep
series CADline
last changed 2003/06/02 13:58

_id fc80
authors Ubbelohde, S. and Humann, C.
year 1998
title Comparative Evaluation of Four Daylighting Software Programs
source 1998 ACEEE Summer Study on Energy Efficiency in Buildings Proceedings. American Council for an Energy-Efficient Economy
summary By the mid-1980's, a number of software packages were under development to predict daylighting performance in buildings, in particular illumination levels in daylighted spaces. An evaluation in 1988 by Ubbelohde et al. demonstrated that none of the software then available was capable of predicting the simplest of real daylighting designs. In the last ten years computer capabilities have evolved rapidly and we have four major packages widely available in the United States. This paper presents a comparative evaluation from the perspective of building and daylighting design practice. A contemporary building completed in 1993 was used as a base case for evaluation. We present the results from field measurements, software predictions and physical modeling as a basis for discussing the capabilities of the software packages in architectural design practice. We found the current software packages far more powerful and nuanced in their ability to predict daylight than previously. Some can accurately predict quantitative daylight performance under varying sky conditions and produce handsome and accurate visualizations of the space. The programs differ significantly, however, in their ease of use, modeling basis and the emphasis between quantitative predictions and visualization in the output.
series other
last changed 2003/04/23 15:50

_id cf2011_p170
id cf2011_p170
authors Barros, Mário; Duarte José, Chaparro Bruno
year 2011
title Thonet Chairs Design Grammar: a Step Towards the Mass Customization of Furniture
source Computer Aided Architectural Design Futures 2011 [Proceedings of the 14th International Conference on Computer Aided Architectural Design Futures / ISBN 9782874561429] Liege (Belgium) 4-8 July 2011, pp. 181-200.
summary The paper presents the first phase of research currently under development that is focused on encoding Thonet design style into a generative design system using a shape grammar. The ultimate goal of the work is the design and production of customizable chairs using computer assisted tools, establishing a feasible practical model of the paradigm of mass customization (Davis, 1987). The current research step encompasses the following three steps: (1) codification of the rules describing Thonet design style into a shape grammar; (2) implementing the grammar into a computer tool as parametric design; and (3) rapid prototyping of customized chair designs within the style. Future phases will address the transformation of the Thonet’s grammar to create a new style and the production of real chair designs in this style using computer aided manufacturing. Beginning in the 1830’s, Austrian furniture designer Michael Thonet began experimenting with forming steam beech, in order to produce lighter furniture using fewer components, when compared with the standards of the time. Using the same construction principles and standardized elements, Thonet produced different chairs designs with a strong formal resemblance, creating his own design language. The kit assembly principle, the reduced number of elements, industrial efficiency, and the modular approach to furniture design as a system of interchangeable elements that may be used to assemble different objects enable him to become a pioneer of mass production (Noblet, 1993). The most paradigmatic example of the described vision of furniture design is the chair No. 14 produced in 1858, composed of six structural elements. Due to its simplicity, lightness, ability to be stored in flat and cubic packaging for individual of collective transportation, respectively, No. 14 became one of the most sold chairs worldwide, and it is still in production nowadays. Iconic examples of mass production are formally studied to provide insights to mass customization studies. The study of the shape grammar for the generation of Thonet chairs aimed to ensure rules that would make possible the reproduction of the selected corpus, as well as allow for the generation of new chairs within the developed grammar. Due to the wide variety of Thonet chairs, six chairs were randomly chosen to infer the grammar and then this was fine tuned by checking whether it could account for the generation of other designs not in the original corpus. Shape grammars (Stiny and Gips, 1972) have been used with sucesss both in the analysis as in the synthesis of designs at different scales, from product design to building and urban design. In particular, the use of shape grammars has been efficient in the characterization of objects’ styles and in the generation of new designs within the analyzed style, and it makes design rules amenable to computers implementation (Duarte, 2005). The literature includes one other example of a grammar for chair design by Knight (1980). In the second step of the current research phase, the outlined shape grammar was implemented into a computer program, to assist the designer in conceiving and producing customized chairs using a digital design process. This implementation was developed in Catia by converting the grammar into an equivalent parametric design model. In the third phase, physical models of existing and new chair designs were produced using rapid prototyping. The paper describes the grammar, its computer implementation as a parametric model, and the rapid prototyping of physical models. The generative potential of the proposed digital process is discussed in the context of enabling the mass customization of furniture. The role of the furniture designer in the new paradigm and ideas for further work also are discussed.
keywords Thonet; furniture design; chair; digital design process; parametric design; shape grammar
series CAAD Futures
email
last changed 2012/02/11 19:21

_id 8629
authors Barzilay, Amos
year 1980
title Human Problem Solving on Master Mind
source Carnegie Mellon University
summary The purpose of this work is to analyze the task of playing Master Mind and to examine subjects behaviors on solving that task. The methods and the ideas that are used in the work are the same found in the references for other tasks. The author wants to show that those ideas and methods can be used for that specific task as well. In other words, subjects behave in such a domain as an information processing system. [includes bibliography]
keywords Psychology, Problem Solving
series CADline
last changed 1999/02/15 15:10

_id a587
authors Cohen, Elaine, Lyche, Tom and Riesenfeld, Richard F.
year 1980
title Discrete B-Splines and Subdivision Techniques in Computer-Aided Geometric Design and Computer Graphics
source computer Graphics and Image Processing. October, 1980. vol. 14: pp. 87- 111 : ill. includes bibliography
summary The relevant theory of discrete B-splines with associated new algorithms is extended to provide a framework for understanding and implementing general subdivision schemes for nonuniform B-splines. The new derived polygon corresponding to an arbitrary refinement of the knot vector for an existing B-spline curve, including multiplicities, is shown to be formed by successive evaluations of the discrete B-spline defined by the original vertices, the original knot vector, and the refined knot vector. Existing subdivision algorithms can be seen as proper special cases. General subdivision has widespread applications in computer-aided geometric design, computer graphics, and numerical analysis. The new algorithms resulting from the new theory lead to a unification of the display model, the analysis model, and other needed models into a single geometric model from which other necessary models are easily derived. New sample algorithms for interference calculation, contouring, surface rendering, and other important calculations are presented
keywords computational geometry, theory, algorithms, computer graphics, B-splines, curved surfaces
series CADline
last changed 2003/06/02 13:58

_id 7f71
authors De Landa, Manuel
year 1998
title Meshworks, Hierarchies and Interfaces
source Virtual Dimension Ed. Beckmann, J., Princeton Architectural Press, NewYork
summary The world of interface design is today undergoing dramatic changes which in their impact promise to rival those brought about by the use of the point-and-click graphical interfaces popularized by the Macintosh in the early 1980's. The new concepts and metaphors which are aiming to replace the familiar desk-top metaphor all revolve around the notion of semi-autonomous, semi-intelligent software agents. To be sure, different researchers and commercial companies have divergent conceptions of what these agents should be capable of, and how they should interact with computer users. But whether one aims to give these software creatures the ability to learn about the users habits, as in the non-commercial research performed at MIT autonomous agents group, or to endow them with the ability to perform transactions in the users name, as in the commercial products pioneered by General Magic, the basic thrust seems to be in the direction of giving software programs more autonomy in their decision-making capabilities.
series other
last changed 2003/04/23 15:50

_id 076e
authors Ennis, G. and Lindsay, M.
year 1999
title VRML Possibilities: The evolution of the Glasgow Model
source Proceedings of International Conference on Virtual Systems and MultiMedia. University of Abertay. Dundee
summary During the 1980's, ABACUS, a research unit at the University of Strathclyde developed an interest in the ability to model and manipulate large geometrical databases of urban topography. Initially, this interest lay solely in the ability to source, capture and store the relevant data. However, once constructed, these models proved genuinely useful to a wide range of users and there was soon a demand for more functionality relating to the manipulation not just of the graphics, but also the range of urban attributes. Although a number of improvements were implemented there were drawbacks to the wide adoption of the software produced. The problems were almost all due to deficiencies in the then current hardware and software system available to the professions, and although this strand of research continued to be pursued, most of the development had to be focused on research applications and deployment. However, the recent advent of the Virtual Reality Modelling Language (VRML) standards have rekindled interest in this field since this language enables many of the issues that have proved problematic in the past to be addressed and solved. The potential now exists to provide wide access to large scale urban models. This paper focuses on the application of VRML as applied to the 'Glasgow Model'.
series other
email
last changed 2003/04/23 15:50

_id 6a59
authors Franklin, Randolph
year 1980
title A Linear Time Exact Hidden Surface Algorithm
source SIGGRAPH '80 Conference Proceedings. July, 1980. vol. 14 ; no. 3: pp. 117-133 : ill. includes bibliography
summary This Paper presents a new hidden surface algorithm. Its output is the set of the visible pieces of edges and faces, and is as accurate as the arithmetic precision of the computer. Thus calculating the hidden surfaces for a higher resolution device takes no more time. If the faces are independently and identically distributed, then the execution time is linear in the number of faces. In particular, the execution time does not increase with the depth complexity. This algorithm overlays a grid on the screen whose fineness depends on the number and size of the faces. Edges and faces are sorted into grid cells. Only objects in the same cell can intersect or hide each other. Also, if a face completely covers a cell then nothing behind it in the cell is relevant. Three programs have tested this algorithm. The first verified the variable grid concept on 50,000 intersecting edges. The second verified the linear time, fast speed, and irrelevance of depth complexity for hidden lines on 10,000 spheres. This also tested depth complexities up to 30, and showed that perspective scenes with the farther objects smaller are even faster to calculate. The third verified this for hidden surfaces on 3,000 squares
keywords hidden surfaces, algorithms, hidden lines, variables, grids, computer graphics, programming
series CADline
last changed 2003/06/02 13:58

_id b190
authors Goldberg, Adele and Robson, David
year 1983
title Smalltalk-80: The language and its implementation
source New York, NY: Addison Wesley Co
summary Smalltalk-80 is the classic standard Smalltalk language as described in Smalltalk-80: The Language and Its Implementation by Goldberg and Robson. This book is commonly called "the Blue Book". Squeak implements the dialect of Smalltalk described in this book, but has a different implementation. Overview of the Smalltalk Language Smalltalk is a general purpose, high level programming language. It was the first original "pure" object oriented language, but not the first to use the object oriented concept, which is credited to Simula 67. The explosive growth of Object Oriented Programming (OOP) technologies began in the early 1980's, with Smalltalk's introduction. Behind it was the idea that the individual human user should be the most important component of any computing system, and that programming should be a natural extension of thinking, and also a dynamic and evolutionary process consistent with the model of human learning activity. In Smalltalk, these ideas are embodied in a framework for human-computer communication. In a sense, Smalltalk is yet another language like C and Pascal, and programs can be written in Smalltalk that have the look and feel of such conventional languages. The difference lies * in the amount of code that can be reduced, * less cryptic syntax, * and code that is easier to handle for application maintenance and enhancement. But Smalltalk's most powerful feature is easy code reuse. Smalltalk makes reuse of programs, routines, and subroutines (methods) far easier. Though procedural languages allow reuse too, it is harder to do, and much easier to cheat. It is no surprise that Smalltalk is relatively easy to learn, mainly due to its simple syntax and semantics, as well as few concepts. Objects, classes, messages, and methods form the basis of programming in Smalltalk. The general methodology to use Smalltalk The notion of human-computer interface also results in Smalltalk promoting the development of safer systems. Errors in Smalltalk may be viewed as objects telling users that confusion exists as to how to perform a desired function.
series other
last changed 2003/04/23 15:14

_id ga9809
id ga9809
authors Kälviäinen, Mirja
year 1998
title The ideological basis of generative expression in design
source International Conference on Generative Art
summary This paper will discuss issues concerning the design ideology supporting the use and development of generative design. This design ideology is based on the unique qualities of craft production and on the forms or ideas from nature or the natural characteristics of materials. The main ideology presented here is the ideology of the 1980´s art craft production in Finland. It is connected with the general Finnish design ideology and with the design ideology of other western countries. The ideology for these professions is based on the common background of design principles stated in 19th century England. The early principles developed through the Arts and Crafts tradition which had a great impact on design thinking in Europe and in the United States. The strong continuity of this design ideology from 19th century England to the present computerized age can be detected. The application of these design principles through different eras shows the difference in the interpretations and in the permission of natural decorative forms. The ideology of the 1980ïs art craft in Finland supports the ideas and fulfilment of generative design in many ways. The reasons often given as the basis for making generative design with computers are in very many respects the same as the ideology for art craft. In Finland there is a strong connection between art craft and design ideology. The characteristics of craft have often been seen as the basis for industrial design skills. The main themes in the ideology of the 1980´s art craft in Finland can be compared to the ideas of generative design. The main issues in which the generative approach reflects a distinctive ideological thinking are: Way of Life: The work is the communication of the maker´s inner ideas. The concrete relationship with the environment, personality, uniqueness, communication, visionary qualities, development and growth of the maker are important. The experiments serve as a media for learning. Taste and Aesthetic Education: The real love affair is created by the non living object with the help of memories and thought. At their best objects create the basis in their stability and communication for durable human relationships. People have warm relationships especially with handmade products in which they can detect unique qualities and the feeling that the product has been made solely for them. Counter-culture: The aim of the work is to produce alternatives for technoburocracy and mechanical production and to bring subjective and unique experiences into the customerïs monotonious life. This ideology rejects the usual standardized mass production of our times. Mythical character: There is a metamorphosis in the birth of the product. In many ways the design process is about birth and growth. The creative process is a development story of the maker. The complexity of communication is the expression of the moments that have been lived. If you can sense the process of making in the product it makes it more real and nearer to life. Each piece of wood has its own beauty. Before you can work with it you must find the deep soul of its quality. The distinctive traits of the material, technique and the object are an essential part of the metamorphosis which brings the product into life. The form is not only for formïs sake but for other purposes, too. You cannot find loose forms in nature. Products have their beginnings in the material and are a part of the nature. This art craft ideology that supports the ideas of generative design can be applied either to the hand made crafts production or to the production exploiting new technology. The unique characteristics of craft and the expression of the material based development are a way to broaden the expression and forms of industrial products. However, for a crafts person it is not meaningful to fill the world with objects. In generative, computer based production this is possible. But maybe the production of unique pieces is still slower and makes the industrial production in that sense more ecological. People will be more attached to personal and unique objects, and thus the life cycle of the objects produced will be longer.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id cdc2008_243
id cdc2008_243
authors Loukissas, Yanni
year 2008
title Keepers of the Geometry: Architects in a Culture of Simulation
source First International Conference on Critical Digital: What Matters(s)? - 18-19 April 2008, Harvard University Graduate School of Design, Cambridge (USA), pp. 243-244
summary “Why do we have to change? We’ve been building buildings for years without CATIA?” Roger Norfleet, a practicing architect in his thirties poses this question to Tim Quix, a generation older and an expert in CATIA, a computer-aided design tool developed by Dassault Systemes in the early 1980’s for use by aerospace engineers. It is 2005 and CATIA has just come into use at Paul Morris Associates, the thirty-person architecture firm where Norfleet works; he is struggling with what it will mean for him, for his firm, for his profession. Computer-aided design is about creativity, but also about jurisdiction, about who controls the design process. In Architecture: The Story of Practice, Architectural theorist Dana Cuff writes that each generation of architects is educated to understand what constitutes a creative act and who in the system of their profession is empowered to use it and at what time. Creativity is socially constructed and Norfleet is coming of age as an architect in a time of technological but also social transition. He must come to terms with the increasingly complex computeraided design tools that have changed both creativity and the rules by which it can operate. In today’s practices, architects use computer-aided design software to produce threedimensional geometric models. Sometimes they use off-the-shelf commercial software like CATIA, sometimes they customize this software through plug-ins and macros, sometimes they work with software that they have themselves programmed. And yet, conforming to Larson’s ideas that they claim the higher ground by identifying with art and not with science, contemporary architects do not often use the term “simulation.” Rather, they have held onto traditional terms such as “modeling” to describe the buzz of new activity with digital technology. But whether or not they use the term, simulation is creating new architectural identities and transforming relationships among a range of design collaborators: masters and apprentices, students and teachers, technical experts and virtuoso programmers. These days, constructing an identity as an architect requires that one define oneself in relation to simulation. Case studies, primarily from two architectural firms, illustrate the transformation of traditional relationships, in particular that of master and apprentice, and the emergence of new roles, including a new professional identity, “keeper of the geometry,” defined by the fusion of person and machine. Like any profession, architecture may be seen as a system in flux. However, with their new roles and relationships, architects are learning that the fight for professional jurisdiction is increasingly for jurisdiction over simulation. Computer-aided design is changing professional patterns of production in architecture, the very way in which professionals compete with each other by making new claims to knowledge. Even today, employees at Paul Morris squabble about the role that simulation software should play in the office. Among other things, they fight about the role it should play in promotion and firm hierarchy. They bicker about the selection of new simulation software, knowing that choosing software implies greater power for those who are expert in it. Architects and their collaborators are in a continual struggle to define the creative roles that can bring them professional acceptance and greater control over design. New technologies for computer-aided design do not change this reality, they become players in it.
email
last changed 2009/01/07 08:05

_id ascaad2007_019
id ascaad2007_019
authors Morsy, S.M.
year 2007
title A Social Approach to Intelligent Buildings
source Em‘body’ing Virtual Architecture: The Third International Conference of the Arab Society for Computer Aided Architectural Design (ASCAAD 2007), 28-30 November 2007, Alexandria, Egypt, pp. 229-242
summary Intelligent buildings were a hot topic since the 1980's. The vision was to share building telecommunication backbones with energy management systems, fire alarms, security systems and even office automation. The benefits were touted to include cost savings for cabling and long term maintenance, plus a future potential for interoperability between systems. The potential within these concepts and the surrounding technology is vast, and lives of users are changing from the effects of Intelligent Buildings developments on their living and working environments. In other words, the main aim of building intelligent building is merely a technological trend. But using intelligent buildings has wider effects on users than intended. Neglecting the human aspects of the users may be considered as a shortcoming. This paper considers the causes of this new trend in architecture in a larger social context, from which the rationalism of building intelligent buildings must arise. This will be done through exploring the concepts and applications of intelligent buildings, showing how all of them affect human life in such many ways.
series ASCAAD
email
last changed 2008/01/21 22:00

_id 48db
authors Proctor, George
year 2001
title CADD Curriculum - The Issue of Visual Acuity
doi https://doi.org/10.52842/conf.ecaade.2001.192
source Architectural Information Management [19th eCAADe Conference Proceedings / ISBN 0-9523687-8-1] Helsinki (Finland) 29-31 August 2001, pp. 192-200
summary Design educators attempt to train the eyes and minds of students to see and comprehend the world around them with the intention of preparing those students to become good designers, critical thinkers and ultimately responsible architects. Over the last eight years we have been developing the digital media curriculum of our architecture program with these fundamental values. We have built digital media use and instruction on the foundation of our program which has historically been based in physical model making. Digital modeling has gradually replaced the capacity of physical models as an analytical and thinking tool, and as a communication and presentation device. The first year of our program provides a foundation and introduction to 2d and 3d design and composition, the second year explores larger buildings and history, the third year explores building systems and structure through design studies of public buildings, fourth year explores urbanism, theory and technology through topic studios and, during the fifth year students complete a capstone project. Digital media and CADD have and are being synchronized with the existing NAAB accredited regimen while also allowing for alternative career options for students. Given our location in the Los Angeles region, many students with a strong background in digital media have gone on to jobs in video game design and the movie industry. Clearly there is much a student of architecture must learn to attain a level of professional competency. A capacity to think visually is one of those skills and is arguably a skill that distinguishes members of the visual arts (including Architecture) from other disciplines. From a web search of information posted by the American Academy of Opthamology, Visual Acuity is defined as an ability to discriminate fine details when looking at something and is often measured with the Snellen Eye Chart (the 20/20 eye test). In the context of this paper visual acuity refers to a subject’s capacity to discriminate useful abstractions in a visual field for the purposes of Visual Thinking- problem solving through seeing (Arnheim, 1969, Laseau 1980, Hoffman 1998). The growing use of digital media and the expanding ability to assemble design ideas and images through point-and-click methods makes the cultivation and development of visual skills all the more important to today’s crop of young architects. The advent of digital media also brings into question the traditional, static 2d methods used to build visual skills in a design education instead of promoting active 3d methods for teaching, learning and developing visual skills. Interactive digital movies provide an excellent platform for promoting visual acuity, and correlating the innate mechanisms of visual perception with the abstractions and notational systems used in professional discourse. In the context of this paper, pedagogy for building visual acuity is being considered with regard to perception of the real world, for example the visual survey of an environment, a site or a street scene and how that visual survey works in conjunction with practice.
keywords Curriculum, Seeing, Abstracting, Notation
series eCAADe
email
last changed 2022/06/07 08:00

_id 5d5a
authors Richens, P.
year 1980
title The Design of a General Drafting System
source CAD80, 4th International Conference and Exhibition of Computers in Design Engineering, Brighton
series other
email
more http://www.arct.cam.ac.uk/research/pubs/
last changed 2000/03/05 19:05

_id 44a3
authors Shen, Tie Gang
year 1996
title Today’s CAAD in China
doi https://doi.org/10.52842/conf.caadria.1996.249
source CAADRIA ‘96 [Proceedings of The First Conference on Computer Aided Architectural Design Research in Asia / ISBN 9627-75-703-9] Hong Kong (Hong Kong) 25-27 April 1996, pp. 249-252
summary The research and application of CAAD in China was in about the 1980’s. Though a lot of success has been achieved in these ten years or so, the improvement is still relatively minor when compared with some other speciality. In fact there are only very few architects who are really making use of CAAD. In this paper, I would like to analyze the present situation and countermeasure CAAD in China.
series CAADRIA
last changed 2022/06/07 07:57

_id acadia21_76
id acadia21_76
authors Smith, Rebecca
year 2021
title Passive Listening and Evidence Collection
doi https://doi.org/10.52842/conf.acadia.2021.076
source ACADIA 2021: Realignments: Toward Critical Computation [Proceedings of the 41st Annual Conference of the Association of Computer Aided Design in Architecture (ACADIA) ISBN 979-8-986-08056-7]. Online and Global. 3-6 November 2021. edited by B. Bogosian, K. Dörfler, B. Farahi, J. Garcia del Castillo y López, J. Grant, V. Noel, S. Parascho, and J. Scott. 76-81.
summary In this paper, I present the commercial, urban-scale gunshot detection system ShotSpotter in contrast with a range of ecological sensing examples which monitor animal vocalizations. Gunshot detection sensors are used to alert law enforcement that a gunshot has occurred and to collect evidence. They are intertwined with processes of criminalization, in which the individual, rather than the collective, is targeted for punishment. Ecological sensors are used as a “passive” practice of information gathering which seeks to understand the health of a given ecosystem through monitoring population demographics, and to document the collective harms of anthropogenic change (Stowell and Sueur 2020). In both examples, the ability of sensing infrastructures to “join up and speed up” (Gabrys 2019, 1) is increasing with the use of machine learning to identify patterns and objects: a new form of expertise through which the differential agendas of these systems are implemented and made visible. I trace the differential agendas of these systems as they manifest through varied components: the spatial distribution of hardware in the existing urban environment and / or landscape; the software and other informational processes that organize and translate the data; the visualization of acoustical sensing data; the commercial factors surrounding the production of material components; and the apps, platforms, and other forms of media through which information is made available to different stakeholders. I take an interpretive and qualitative approach to the analysis of these systems as cultural artifacts (Winner 1980), to demonstrate how the political and social stakes of the technology are embedded throughout them.
series ACADIA
type paper
email
last changed 2023/10/22 12:06

_id 9fcb
authors Steele, Guy Lewis
year 1980
title The Definition and Implementation of a Computer Programming Language Based Constraints
source MIT - AITR-595
summary The constraint paradigm is a model of computation in which values are deduced whenever possible, under the limitation that deductions be local in a certain sense. One may visualize a constraint 'program' as a network of devices connected by wires. Data values may flow along the wires, and computation is performed by the devices. A device computes using only locally available information (with a few exceptions), and places newly derived values on other, locally attached wires. In this way computed values are propagated. An advantage of the constraint paradigm (not unique to it) is that a single relationship can be used in more than one direction. The connections to a device are not labelled as inputs and outputs; a device will compute with whatever values are available, and produce as many new values as it can. General theorem provers are capable of such behavior, but tend to suffer from combinatorial explosion; it is not usually useful to derive all the possible consequences of a set of hypotheses. The constraint paradigm places a certain kind of limitation on the deduction process. The limitations imposed by the constraint paradigm are not the only one possible. It is argued, however, that they are restrictive enough to forestall combinatorial explosion in many interesting computational situations, yet permissive enough to allow useful computations in practical situations. Moreover, the paradigm is intuitive: It is easy to visualize the computational effects of these particular limitations, and the paradigm is a natural way of expressing programs for certain applications, in particular relationships arising in computer-aided design. A number of implementations of constraint-based programming languages are presented. A progression of ever more powerful languages is described, complete implementations are presented and design difficulties and alternatives are discussed. The goal approached, though not quite reached, is a complete programming system which will implicitly support the constraint paradigm to the same extent that LISP, say, supports automatic storage management.
series thesis:PhD
email
more ftp://publications.ai.mit.edu/ai-publications/pdf/AITR-595.pdf
last changed 2003/02/12 22:37

_id e825
authors Baybars, Ilker and Eastman, Charles M.
year 1980
title Enumerating Architectural Arrangements by Generating Their Underlying Graphs
source Environment and Planning B. 1980. vol. 7: pp. 289- 310 : ill. includes bibliography. -- See also 'Enumerating Architectural Arrangements: Comment on a Recent Paper by Baybars and Eastman' by C.F. Earl
summary One mathematical correspondence to the partitioning of the plane is a Weighted Plane Graph (WPG). This paper first focuses on the systematic generation of WPGs, in a fashion similar to crystal growth. During this process, the WPGs are represented by adjacency matrices. The authors, thus, present a method for embedding the WPG in the plane, given its adjacency matrix. These graphs can, then, be mapped into floor plans. The common practice here is the use of the `geometric dual' of a WPG. The authors propose, instead, the use of the `Pseudogeometric dual' of a WPG directly to translate (part of) a design brief into alternative spatial layouts. Also discussed is the ability to create courtyards and/or circulation spaces given a specific WPG, without increasing the size of the problem
keywords enumeration, architecture, floor plans, graphs, design process, automation, algorithms, space allocation, CAD
series CADline
email
last changed 2003/05/17 10:15

_id acadia07_164
id acadia07_164
authors Diniz, Nancy; Turner, Alasdair
year 2007
title Towards a Living Architecture
doi https://doi.org/10.52842/conf.acadia.2007.164
source Expanding Bodies: Art • Cities• Environment [Proceedings of the 27th Annual Conference of the Association for Computer Aided Design in Architecture / ISBN 978-0-9780978-6-8] Halifax (Nova Scotia) 1-7 October 2007, 164-173
summary Interaction is the latest currency in architecture, as responsive components are now reacting to the inhabitant of the space. These components are designed and installed by the architect with a view to the phenomenology of space, where the experience of the environment is previewed and pre-constructed before it is translated into the conception of the space. However, this traditional approach to new technology leaves no scope for the architecture to be alive in and of itself, and thus the installed piece quickly becomes just that—an installation: isolated and uncontained by its environment. In this paper, we argue that a way to approach a responsive architecture is to design for a piece that is truly living, and in order to propose a living architecture first we need to understand what the architecture of a living system is. This paper suggests a conceptual framework based on the theory of Autopoiesis in order to create a “self-producing” system through an experiment entitled, “The Life of a Wall” (Maturana and Varela 1980). The wall has a responsive membrane controlled by a genetic algorithm that reconfigures its behaviour and learns to adapt itself continually to the evolutionary properties of the environment, thus becoming a situated, living piece.
series ACADIA
email
last changed 2022/06/07 07:55

For more results click below:

this is page 0show page 1show page 2show page 3show page 4HOMELOGIN (you are user _anon_575438 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002