CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 10 of 10

_id cdc2008_243
id cdc2008_243
authors Loukissas, Yanni
year 2008
title Keepers of the Geometry: Architects in a Culture of Simulation
source First International Conference on Critical Digital: What Matters(s)? - 18-19 April 2008, Harvard University Graduate School of Design, Cambridge (USA), pp. 243-244
summary “Why do we have to change? We’ve been building buildings for years without CATIA?” Roger Norfleet, a practicing architect in his thirties poses this question to Tim Quix, a generation older and an expert in CATIA, a computer-aided design tool developed by Dassault Systemes in the early 1980’s for use by aerospace engineers. It is 2005 and CATIA has just come into use at Paul Morris Associates, the thirty-person architecture firm where Norfleet works; he is struggling with what it will mean for him, for his firm, for his profession. Computer-aided design is about creativity, but also about jurisdiction, about who controls the design process. In Architecture: The Story of Practice, Architectural theorist Dana Cuff writes that each generation of architects is educated to understand what constitutes a creative act and who in the system of their profession is empowered to use it and at what time. Creativity is socially constructed and Norfleet is coming of age as an architect in a time of technological but also social transition. He must come to terms with the increasingly complex computeraided design tools that have changed both creativity and the rules by which it can operate. In today’s practices, architects use computer-aided design software to produce threedimensional geometric models. Sometimes they use off-the-shelf commercial software like CATIA, sometimes they customize this software through plug-ins and macros, sometimes they work with software that they have themselves programmed. And yet, conforming to Larson’s ideas that they claim the higher ground by identifying with art and not with science, contemporary architects do not often use the term “simulation.” Rather, they have held onto traditional terms such as “modeling” to describe the buzz of new activity with digital technology. But whether or not they use the term, simulation is creating new architectural identities and transforming relationships among a range of design collaborators: masters and apprentices, students and teachers, technical experts and virtuoso programmers. These days, constructing an identity as an architect requires that one define oneself in relation to simulation. Case studies, primarily from two architectural firms, illustrate the transformation of traditional relationships, in particular that of master and apprentice, and the emergence of new roles, including a new professional identity, “keeper of the geometry,” defined by the fusion of person and machine. Like any profession, architecture may be seen as a system in flux. However, with their new roles and relationships, architects are learning that the fight for professional jurisdiction is increasingly for jurisdiction over simulation. Computer-aided design is changing professional patterns of production in architecture, the very way in which professionals compete with each other by making new claims to knowledge. Even today, employees at Paul Morris squabble about the role that simulation software should play in the office. Among other things, they fight about the role it should play in promotion and firm hierarchy. They bicker about the selection of new simulation software, knowing that choosing software implies greater power for those who are expert in it. Architects and their collaborators are in a continual struggle to define the creative roles that can bring them professional acceptance and greater control over design. New technologies for computer-aided design do not change this reality, they become players in it.
email
last changed 2009/01/07 08:05

_id a5c3
authors Er, M.C.
year 1981
title The Relations of the Computation of Fibonnaci Numbers with the Polyphase Sort
source 8 p. Wollongong: Department of Computing Science, University of Wollongong, September, 1981. includes bibliography
summary The theory of polyphase sort has simplified the mathematical derivations of Wilson and Shortt's (1980) algorithm, and offered an intuitive explanation of why Gries and Levin's (1980), and Urbanek's (1980) algorithms work. The computation of order-k Fibonacci numbers is equivalent to moving a window of matrix upwards in a series of ideal distributions
keywords Fibonacci, sorting, mathematics, algorithms
series CADline
last changed 2003/06/02 13:58

_id b04c
authors Goerger, S., Darken, R., Boyd, M., Gagnon, T., Liles, S., Sullivan, J. and Lawson, J.
year 1996
title Spatial Knowledge Acquisition from Maps and Virtual Environments in Complex Architectural Space
source Proc. 16 th Applied Behavioral Sciences Symposium, 22-23 April, U.S. Airforce Academy, Colorado Springs, CO., 1996, 6-10
summary It has often been suggested that due to its inherent spatial nature, a virtual environment (VE) might be a powerful tool for spatial knowledge acquisition of a real environment, as opposed to the use of maps or some other two-dimensional, symbolic medium. While interesting from a psychological point of view, a study of the use of a VE in lieu of a map seems nonsensical from a practical point of view. Why would the use of a VE preclude the use of a map? The more interesting investigation would be of the value added of the VE when used with a map. If the VE could be shown to substantially improve navigation performance, then there might be a case for its use as a training tool. If not, then we have to assume that maps continue to be the best spatial knowledge acquisition tool available. An experiment was conducted at the Naval Postgraduate School to determine if the use of an interactive, three-dimensional virtual environment would enhance spatial knowledge acquisition of a complex architectural space when used in conjunction with floor plan diagrams. There has been significant interest in this research area of late. Witmer, Bailey, and Knerr (1995) showed that a VE was useful in acquiring route knowledge of a complex building. Route knowledge is defined as the procedural knowledge required to successfully traverse paths between distant locations (Golledge, 1991). Configurational (or survey) knowledge is the highest level of spatial knowledge and represents a map-like internal encoding of the environment (Thorndyke, 1980). The Witmer study could not confirm if configurational knowledge was being acquired. Also, no comparison was made to a map-only condition, which we felt is the most obvious alternative. Comparisons were made only to a real world condition and a symbolic condition where the route is presented verbally.
series other
last changed 2003/04/23 15:50

_id 6f57
authors Searle, John R.
year 1980
title Minds, Brains, and Programs
source The Behavioral and Brain Sciences. Cambridge University Press., 1980. vol. 3: pp. 417-457. includes bibliography
summary This article can be viewed as an attempt to explore the consequences of two propositions: (1) Intentionallity in human beings (and animals) is a product of causal features of the brain. The author assumes this is an empirical fact about the actual causal relations between mental processes and brains. It says simply that certain brain processes are sufficient for intentionallity. (2) Instantiating a computer program is never by itself a sufficient condition of intentionallity. The main argument of this paper is directed at establishing this claim. The form of the argument is to show how a human agent could instantiate the program and still not have the relevant intentionallity. These two propositions have the following consequences: (3) The explanation of how the brain produces intentionallity cannot be that it does it by instantiating a computer program. This is a strict logical consequence of 1 and 2. (4) Any mechanism capable of producing intentionallity must have causal powers equal to those of the brain. This is meant to be a trivial consequence of 1. (5) Any attempt literally to create intentionallity artificially (strong AI) could not succeed just by designing programs but would have to duplicate the causal powers of the human brain. This follows from 2 and 4. 'Could a machine think?' On the argument advanced here only a machine could think, and only very special kinds of machines, namely brains and machines withÔ h) 0*0*0*°° ÔŒ internal causal powers equivalent to those of brains
keywords And that is why strong AI has little to tell us about thinking, since
series CADline
last changed 2003/06/02 10:24

_id ascaad2007_019
id ascaad2007_019
authors Morsy, S.M.
year 2007
title A Social Approach to Intelligent Buildings
source Em‘body’ing Virtual Architecture: The Third International Conference of the Arab Society for Computer Aided Architectural Design (ASCAAD 2007), 28-30 November 2007, Alexandria, Egypt, pp. 229-242
summary Intelligent buildings were a hot topic since the 1980's. The vision was to share building telecommunication backbones with energy management systems, fire alarms, security systems and even office automation. The benefits were touted to include cost savings for cabling and long term maintenance, plus a future potential for interoperability between systems. The potential within these concepts and the surrounding technology is vast, and lives of users are changing from the effects of Intelligent Buildings developments on their living and working environments. In other words, the main aim of building intelligent building is merely a technological trend. But using intelligent buildings has wider effects on users than intended. Neglecting the human aspects of the users may be considered as a shortcoming. This paper considers the causes of this new trend in architecture in a larger social context, from which the rationalism of building intelligent buildings must arise. This will be done through exploring the concepts and applications of intelligent buildings, showing how all of them affect human life in such many ways.
series ASCAAD
email
last changed 2008/01/21 22:00

_id 48db
authors Proctor, George
year 2001
title CADD Curriculum - The Issue of Visual Acuity
doi https://doi.org/10.52842/conf.ecaade.2001.192
source Architectural Information Management [19th eCAADe Conference Proceedings / ISBN 0-9523687-8-1] Helsinki (Finland) 29-31 August 2001, pp. 192-200
summary Design educators attempt to train the eyes and minds of students to see and comprehend the world around them with the intention of preparing those students to become good designers, critical thinkers and ultimately responsible architects. Over the last eight years we have been developing the digital media curriculum of our architecture program with these fundamental values. We have built digital media use and instruction on the foundation of our program which has historically been based in physical model making. Digital modeling has gradually replaced the capacity of physical models as an analytical and thinking tool, and as a communication and presentation device. The first year of our program provides a foundation and introduction to 2d and 3d design and composition, the second year explores larger buildings and history, the third year explores building systems and structure through design studies of public buildings, fourth year explores urbanism, theory and technology through topic studios and, during the fifth year students complete a capstone project. Digital media and CADD have and are being synchronized with the existing NAAB accredited regimen while also allowing for alternative career options for students. Given our location in the Los Angeles region, many students with a strong background in digital media have gone on to jobs in video game design and the movie industry. Clearly there is much a student of architecture must learn to attain a level of professional competency. A capacity to think visually is one of those skills and is arguably a skill that distinguishes members of the visual arts (including Architecture) from other disciplines. From a web search of information posted by the American Academy of Opthamology, Visual Acuity is defined as an ability to discriminate fine details when looking at something and is often measured with the Snellen Eye Chart (the 20/20 eye test). In the context of this paper visual acuity refers to a subject’s capacity to discriminate useful abstractions in a visual field for the purposes of Visual Thinking- problem solving through seeing (Arnheim, 1969, Laseau 1980, Hoffman 1998). The growing use of digital media and the expanding ability to assemble design ideas and images through point-and-click methods makes the cultivation and development of visual skills all the more important to today’s crop of young architects. The advent of digital media also brings into question the traditional, static 2d methods used to build visual skills in a design education instead of promoting active 3d methods for teaching, learning and developing visual skills. Interactive digital movies provide an excellent platform for promoting visual acuity, and correlating the innate mechanisms of visual perception with the abstractions and notational systems used in professional discourse. In the context of this paper, pedagogy for building visual acuity is being considered with regard to perception of the real world, for example the visual survey of an environment, a site or a street scene and how that visual survey works in conjunction with practice.
keywords Curriculum, Seeing, Abstracting, Notation
series eCAADe
email
last changed 2022/06/07 08:00

_id fc80
authors Ubbelohde, S. and Humann, C.
year 1998
title Comparative Evaluation of Four Daylighting Software Programs
source 1998 ACEEE Summer Study on Energy Efficiency in Buildings Proceedings. American Council for an Energy-Efficient Economy
summary By the mid-1980's, a number of software packages were under development to predict daylighting performance in buildings, in particular illumination levels in daylighted spaces. An evaluation in 1988 by Ubbelohde et al. demonstrated that none of the software then available was capable of predicting the simplest of real daylighting designs. In the last ten years computer capabilities have evolved rapidly and we have four major packages widely available in the United States. This paper presents a comparative evaluation from the perspective of building and daylighting design practice. A contemporary building completed in 1993 was used as a base case for evaluation. We present the results from field measurements, software predictions and physical modeling as a basis for discussing the capabilities of the software packages in architectural design practice. We found the current software packages far more powerful and nuanced in their ability to predict daylight than previously. Some can accurately predict quantitative daylight performance under varying sky conditions and produce handsome and accurate visualizations of the space. The programs differ significantly, however, in their ease of use, modeling basis and the emphasis between quantitative predictions and visualization in the output.
series other
last changed 2003/04/23 15:50

_id ddss2004_ra-33
id ddss2004_ra-33
authors Diappi, L., P. Bolchim, and M. Buscema
year 2004
title Improved Understanding of Urban Sprawl Using Neural Networks
source Van Leeuwen, J.P. and H.J.P. Timmermans (eds.) Recent Advances in Design & Decision Support Systems in Architecture and Urban Planning, Dordrecht: Kluwer Academic Publishers, ISBN: 14020-2408-8, p. 33-49
summary It is widely accepted that the spatial pattern of settlements is a crucial factor affecting quality of life and environmental sustainability, but few recent studies have attempted to examine the phenomenon of sprawl by modelling the process rather than adopting a descriptive approach. The issue was partly addressed by models of land use and transportation which were mainly developed in the UK and US in the 1970s and 1980s, but the major advances were made in the area of modelling transportation, while very little was achieved in the area of spatial and temporal land use. Models of land use and transportation are well-established tools, based on explicit, exogenouslyformulated rules within a theoretical framework. The new approaches of artificial intelligence, and in particular, systems involving parallel processing, (Neural Networks, Cellular Automata and Multi-Agent Systems) defined by the expression “Neurocomputing”, allow problems to be approached in the reverse, bottom-up, direction by discovering rules, relationships and scenarios from a database. In this article we examine the hypothesis that territorial micro-transformations occur according to a local logic, i.e. according to use, accessibility, the presence of services and conditions of centrality, periphericity or isolation of each territorial “cell” relative to its surroundings. The prediction capabilities of different architectures of supervised Neural networks are implemented to the south Metropolitan area of Milan at two different temporal thresholds and discussed. Starting from data on land use in 1980 and 1994 and by subdividing the area into square cells on an orthogonal grid, the model produces a spatial and functional map of urbanisation in 2008. An implementation of the SOM (Self Organizing Map) processing to the Data Base allows the typologies of transformation to be identified, i.e. the classes of area which are transformed in the same way and which give rise to territorial morphologies; this is an interesting by-product of the approach.
keywords Neural Networks, Self-Organizing Maps, Land-Use Dynamics, Supervised Networks
series DDSS
last changed 2004/07/03 22:13

_id 6a59
authors Franklin, Randolph
year 1980
title A Linear Time Exact Hidden Surface Algorithm
source SIGGRAPH '80 Conference Proceedings. July, 1980. vol. 14 ; no. 3: pp. 117-133 : ill. includes bibliography
summary This Paper presents a new hidden surface algorithm. Its output is the set of the visible pieces of edges and faces, and is as accurate as the arithmetic precision of the computer. Thus calculating the hidden surfaces for a higher resolution device takes no more time. If the faces are independently and identically distributed, then the execution time is linear in the number of faces. In particular, the execution time does not increase with the depth complexity. This algorithm overlays a grid on the screen whose fineness depends on the number and size of the faces. Edges and faces are sorted into grid cells. Only objects in the same cell can intersect or hide each other. Also, if a face completely covers a cell then nothing behind it in the cell is relevant. Three programs have tested this algorithm. The first verified the variable grid concept on 50,000 intersecting edges. The second verified the linear time, fast speed, and irrelevance of depth complexity for hidden lines on 10,000 spheres. This also tested depth complexities up to 30, and showed that perspective scenes with the farther objects smaller are even faster to calculate. The third verified this for hidden surfaces on 3,000 squares
keywords hidden surfaces, algorithms, hidden lines, variables, grids, computer graphics, programming
series CADline
last changed 2003/06/02 13:58

_id acadia21_76
id acadia21_76
authors Smith, Rebecca
year 2021
title Passive Listening and Evidence Collection
doi https://doi.org/10.52842/conf.acadia.2021.076
source ACADIA 2021: Realignments: Toward Critical Computation [Proceedings of the 41st Annual Conference of the Association of Computer Aided Design in Architecture (ACADIA) ISBN 979-8-986-08056-7]. Online and Global. 3-6 November 2021. edited by B. Bogosian, K. Dörfler, B. Farahi, J. Garcia del Castillo y López, J. Grant, V. Noel, S. Parascho, and J. Scott. 76-81.
summary In this paper, I present the commercial, urban-scale gunshot detection system ShotSpotter in contrast with a range of ecological sensing examples which monitor animal vocalizations. Gunshot detection sensors are used to alert law enforcement that a gunshot has occurred and to collect evidence. They are intertwined with processes of criminalization, in which the individual, rather than the collective, is targeted for punishment. Ecological sensors are used as a “passive” practice of information gathering which seeks to understand the health of a given ecosystem through monitoring population demographics, and to document the collective harms of anthropogenic change (Stowell and Sueur 2020). In both examples, the ability of sensing infrastructures to “join up and speed up” (Gabrys 2019, 1) is increasing with the use of machine learning to identify patterns and objects: a new form of expertise through which the differential agendas of these systems are implemented and made visible. I trace the differential agendas of these systems as they manifest through varied components: the spatial distribution of hardware in the existing urban environment and / or landscape; the software and other informational processes that organize and translate the data; the visualization of acoustical sensing data; the commercial factors surrounding the production of material components; and the apps, platforms, and other forms of media through which information is made available to different stakeholders. I take an interpretive and qualitative approach to the analysis of these systems as cultural artifacts (Winner 1980), to demonstrate how the political and social stakes of the technology are embedded throughout them.
series ACADIA
type paper
email
last changed 2023/10/22 12:06

No more hits.

HOMELOGIN (you are user _anon_551267 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002