CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 26

_id ea14
authors Anson, Ed
year 1979
title The Semantics of Graphical Input
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 113- 120. includes bibliography
summary This paper describes the semantics of action, an approach to describing input devices which allow the full utilization of all useful device characteristics and provides a high degree of hardware device independence. Part one discusses the semantics of graphical input device. The second shows how to create hierarchies of devices which provide a large measure of hardware independence. The third part applies these concepts to some typical problems, to demonstrate their completeness
keywords computer graphics, user interface, semantics
series CADline
last changed 1999/02/12 15:07

_id 00f3
authors Baybars, Ilker and Eastman, Charles M.
year 1979
title Generating the Underlying Graphs for Architectural Arrangements
source 10 p. : ill. Pittsburgh: School of Urban and Public Affairs, Carnegie Mellon University, April, 1979. Research report No.79. Includes bibliography
summary The mathematical correspondence to a floorplan is a Metric Planar Graph. Several methods for systematic direct generation of metric planar graphs have been developed including polyominoes, March and Matela and shape grammars. Another approach has been to develop a spatial composition in two separate steps. The first step involves discrete variables, and consists of enumerating a defined set of non-metric planar graphs. The second step involves spatial dimensions, e.g. continuous variables, and maps the graphs onto the Euclidean plane, from which a satisfactory or optimal one is selected. This paper focusses on the latter 2-step process. It presents a general method of solving the first step, that is the exhaustive enumeration of a set of planar graphs. The paper consists of three sections: The first section is an introduction to graph theory. The second section presents the generation of maximal planar graphs. The last section summarizes the presentation and comments on the appropriateness of the method
keywords graphs, floor plans, architecture, design, automation, space allocation
series CADline
email
last changed 2003/05/17 10:15

_id fcd6
authors Berger, S.R.
year 1979
title Artificial Intelligence and its impact on Coimputer-Aided Design
source Design Studies, vol 1, no. 3
summary This paper provides, for readers unfamiliar with the field, an introductory account of research which has been carried out in artificial intelligence. It attempts to distingussh between an artificial intelligence and a conventional computing approach and to assess the future influence of the former on computer-aided design.
series journal paper
last changed 2003/04/23 15:14

_id 6733
authors Bettels, Juergen and Myers, David R.
year 1986
title The PIONS Graphics System
source IEEE Computer Graphics and Applications. July, 1986. vol. 6: pp. 30-38 : col. ill. includes a short bibliography
summary During 1979, CERN began to evaluate how interactive computer graphics displays could aid the analysis of high-energy physics experiments at the new Super Proton Synchrotron collider. This work led to PIONS, a 3D graphics system, which features the ability to store and view hierarchical graphics structures in a directed-acyclic-graph database. It is possible to change the attributes of these structures by making selections on nongraphical information also stored in the database. PIONS is implemented as an object-oriented message-passing system based on SmallTalk design principles. It supports multiple viewing transformations, logical input devices, and 2D and 3D primitives. The design allows full use to be made of display hardware that provides dynamic 3D picture transformation
keywords visualization, computer graphics, database, systems, modeling
series CADline
last changed 2003/06/02 13:58

_id af53
authors Boyer, E. and Mitgang, L.
year 1996
title Building community: a new future for architecture education and practice
source Carnegie Foundation for the Advancement of Teaching
summary Internships, before and after graduation, are the most essential link connecting students to the world of practice. Yet, by all accounts, internship is perhaps the most troubled phase of the continuing education of architects. During this century, as architectural knowledge grew more complex, the apprenticeship system withered away and schools assumed much of the responsibility for preparing architects for practice. However, schools cannot do the whole job. It is widely acknowledged that certain kinds of technical and practical knowledge are best learned in the workplace itself, under the guidance of experienced professionals. All state accrediting boards require a minimum period of internship-usually about three years-before a person is eligible to take the licensing exam. The National Council of Architectural Registration Boards (NCARB) allows students to earn up to two years of work credit prior to acquisition of an accredited degree. The Intern Development Program (IDP), launched by NCARB and the American Institute of Architects in 1979, provides the framework for internship in some forty states. The program was designed to assure that interns receive adequate mentoring, that experiences are well-documented, and that employers and interns allocate enough time to a range of educational and vocational experiences to prepare students for eventual licensure. As the IDP Guidelines state, "The shift from school to office is not a transition from theory to pragmatism. It is a period when theory merges with pragmatism.... It's a time when you: apply your formal education to the daily realities of architectural practice; acquire comprehensive experience in basic practice areas; explore specialized areas of practice; develop professional judgment; continue your formal education in architecture; and refine your career goals." Whatever its accomplishments, however, we found broad consensus that the Intern Development Program has not, by itself, solved the problems of internship. Though we found mutually satisfying internship programs at several of the firms we visited or heard about around the country, at many others interns told us they were not receiving the continuing education and experience they needed. The truth is that architecture has serious, unsolved problems compared with other fields when it comes to supplying on-the-job learning experiences to induct students into the profession on a massive scale. Medicine has teaching hospitals. Beginning teachers work in actual classrooms, supported by school taxes. Law offices are, for the most part, in a better financial position to support young lawyers and pay them living wages. The architecture profession, by contrast, must support a required system of internship prior to licensure in an industry that has neither the financial resources of law or medicine, the stability and public support of teaching, nor a network of locations like hospitals or schools where education and practice can be seamlessly connected. And many employers acknowledged those problems. "The profession has all but undermined the traditional relationship between the profession and the academy," said Neil Frankel, FAIA, executive vice president of Perkins & Will, a multinational firm with offices in New York, Chicago, Washington, and London. "Historically, until the advent of the computer, the profession said, 'Okay, go to school, then we in the profession will teach you what the real world is like.' With the coming of the computer, the profession needed a skill that students had, and has left behind the other responsibilities." One intern told us she had been stuck for months doing relatively menial tasks such as toilet elevations. Another intern at a medium-sized firm told us he had been working sixty to seventy hours per week for a year and a half. "Then my wife had a baby and I 'slacked off' to fifty hours. The partner called me in and I got called on the carpet for not working hard enough." "The whole process of internship is being outmoded by economics," one frustrated intern told us. "There's not the time or the money. There's no conception of people being groomed for careers. The younger staff are chosen for their value as productive workers." "We just don't have the best structure here to use an intern's abilities to their best," said a Mississippi architect. "The people who come out of school are really problems. I lost patience with one intern who was demanding that I switch him to another section so that he could learn what he needed for his IDP. I told him, 'It's not my job to teach you. You are here to produce.'" What steps might help students gain more satisfying work opportunities, both during and after graduation?
series other
last changed 2003/04/23 15:14

_id 819a
authors Brassel, Kurt E. and Fegeas, Robin
year 1979
title An Algorithm for Shading of Regions on Vector Display Devices
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 126- 133 : ill. includes bibliography
summary The display of shaded polygons by line, cross-hatch, and dot patterns on vector devices is a task frequently used in computer graphics and computer cartography. In applications such as the production of shaded maps polygon shading turns out to be critical with respect to time requirements, and the development of efficient algorithms is of importance. Given an arbitrary polygon in the plane without self-crossing edges (simply-connected polygon), the task at hand is to shade this polygon with one or two sets of parallel lines where for each set a shading angle and a line distance are given. The basic concept of this new algorithm is to decompose the polygon into a set of mutually exclusive trapezoids (in special cases triangles) where the parallel edges of the trioxides are parallel to the desired shading lines. These trapezoids and triangles are then shaded in a fast procedure. In its present form the algorithm handles regions with up to 300 islands. Possible extensions include the construction of dash and cross patterns
keywords algorithms, polygons, software, computer graphics, shading, GIS, mapping, drafting, information
series CADline
last changed 2003/06/02 13:58

_id 4435
authors Cheatham, Th.E., Townley, J.A. and Holloway, G.H.
year 1979
title A System for Program Refinement
source 1979. pp. 53-62. includes bibliography
summary The Program Development System (PDS) is a programming environment, an integrated collection of interactive tools that support the process of program definition, testing, and maintenance. The PDS is intended to aid the development of large programs, especially program families whose members must be maintained in synchrony. The system facilitates implementation by stepwise refinement, and it keeps a refinement history that allows program modifications made at a high level of abstraction to be reflected efficiently and automatically in the corresponding low level code. Analysis tools are used both to support program validation and to guide program refinement
keywords user interface, software, systems, programming, tools
series CADline
last changed 2003/06/02 14:41

_id ga0015
id ga0015
authors Daru, R., Vreedenburgh, E. and Scha, R.
year 2000
title Architectural Innovation as an evolutionary process
source International Conference on Generative Art
summary Traditionally in art and architectural history, innovation is treated as a history of ideas of individuals (pioneers), movements and schools. The monograph is in that context one of the most used forms of scientific exercise. History of architecture is then mostly seen as a succession of dominant architectural paradigms imposed by great architectural creators fighting at the beginning against mainstream establishment until they themselves come to be recognised. However, there have been attempts to place architectural innovation and creativity in an evolutionary perspective. Charles Jencks for example, has described the evolution of architectural and art movements according to a diagram inspired by ecological models. Philip Steadman, in his book "The Evolution of Designs. Biological analogy in architecture and the applied arts" (1979), sketches the history of various biological analogies and their impact on architectural theory: the organic, classificatory, anatomical, ecological and Darwinian or evolutionary analogies. This last analogy "explains the design of useful objects and buildings, particularly in primitive society and in the craft tradition, in terms of a sequence of repeated copyings (corresponding to inheritance), with small changes made at each stage ('variations'), which are then subjected to a testing process when the object is put into use ('selection')." However, Steadman has confined his study to a literature survey as the basis of a history of ideas. Since this pioneering work, new developments like Dawkins' concept of memes allow further steps in the field of cultural evolution of architectural innovation. The application of the concept of memes to architectural design has been put forward in a preceding "Generative Art" conference (Daru, 1999), showing its application in a pilot study on the analysis of projects of and by architectural students. This first empirical study is now followed by a study of 'real life' architectural practice. The case taken has a double implication for the evolutionary analogy. It takes a specific architectural innovative concept as a 'meme' and develops the analysis of the trajectory of this meme in the individual context of the designer and at large. At the same time, the architect involved (Eric Vreedenburgh, Archipel Ontwerpers) is knowledgeable about the theory of memetic evolution and is applying a computer tool (called 'Artificial') together with Remko Scha, the authoring computer scientist of the program who collaborates frequently with artists and architects. This case study (the penthouse in Dutch town planning and the application of 'Artificial') shall be discussed in the paper as presented. The theoretical and methodological problems of various models of diffusion of memes shall be discussed and a preliminary model shall be presented as a framework to account for not only Darwinian but also Lamarckian processes, and for individual as well as collective transmission, consumption and creative transformation of memes.
keywords evolutionary design, architectural innovation, memetic diffusion, CAAD, penthouses, Dutch design, creativity, Darwinian and Lamarckian processes
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 6890
authors Eastman, Charles M. and Weiler, Kevin
year 1979
title Geometric Modeling Using the Euler Operators
source 12 p. : ill. Pittsburgh: Institute of Physical Planning, Carnegie Mellon University, February, 1979. includes bibliography
summary A recent advance in the modeling of three-dimensional shapes is the joint development of bounded shape models, capable of representing complete and well-formed arbitrary polyhedra, and operators for manipulating them. Two approaches have been developed thus far in forming bounded shape models: to combine a given fixed set of primitive shapes into other possibly more complex ones using the spatial set operators, and/or to apply lower level operators that define and combine faces, edges, loops and vertices to directly construct a shape. The name that has come to be applied to these latter operators is the Euler operators. This paper offers a description of the Euler operators, in a form expected to be useful for prospective implementers and others wishing to better understand their function and behavior. It includes considerations regarding their specification in terms of being able to completely describe different classes of shapes, how to properly specify them and the extent of their well-formedness, especially in terms of their interaction with geometric operations. Example specifications are provided as well as some useful applications. The Euler operators provide different capabilities from the spatial set operators. An extensible CAD/CAM facility needs them both
keywords Euler operators, boolean operations, CSG, geometric modeling, CAD, CAM, B-rep, solid modeling, theory
series CADline
email
last changed 2003/05/17 10:15

_id e83f
authors Foley, James D., Templeman, James N. and Dastyar, Dara
year 1979
title Some Raster Graphics Extensions to the Core System
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 15-24 : ill. includes bibliography
summary The Core System is a proposed standard subroutine package for line-drawing graphics. Some Core System extensions are presented for use with raster graphics equipment. The extensions, which are upward-compatible with the present Core System, provide filled polygons, display of arrays of pixels, use of a color look-up table, and hidden-surface on hidden- edge removal. The extensions are being integrated into an existing Core System implementation
keywords hardware, computer graphics, standards,
series CADline
last changed 2003/06/02 13:58

_id eb8e
authors Fowler, Robert J. and Little, James J.
year 1979
title Automatic Extraction of Irregular Network Digital Terrain Models
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 199- 207 : ill. includes bibliography
summary For representation of terrain, an efficient alternative to dense grids is the Triangulated Irregular Network (TIN), which represents a surface as a set of non-overlapping contiguous triangular facets, of irregular size and shape. The source of digital terrain data is increasingly dense raster models produced by automated orthophoto machines or by direct sensors such as synthetic aperture radar. A method is described for automatically extracting a TIN model from dense raster data. An initial approximation is constructed by automatically triangulating a set of feature points derived from the raster model. The method works by local incremental refinement of this model by the addition of new points until a uniform approximation of specified tolerance is obtained. Empirical results show that substantial savings in storage can be obtained
keywords GIS, mapping, computational geometry, data structures, mapping, representation, computer graphics, triangulation
series CADline
last changed 2003/06/02 13:58

_id 4517
authors Fuchs, Henry, Kedem, Zvi M. and Naylor, Bruce F.
year 1979
title Predetermining Visibility Priority in 3-D Scenes
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 175-181 : ill. includes bibliography
summary The principal calculation performed by all visible surface algorithms is the determination of the visible polygon at each pixel in the image. Of the many possible speedups and efficiencies found for this problem, only one published algorithm (developed almost a decade ago by a group at General Electric) took advantage of an observation that many visibility calculations could be performed without knowledge of the eventual viewing position and orientation -- once for all possible images. The method is based on a 'potential obscuration' relation between polygons in the simulated environment. Unfortunately, the method worked only for certain objects; unmanageable objects had to be manually (and expertly!) subdivided into manageable pieces. Described in this paper is a solution to this problem which allows substantial a-priori visibility determination for all possible objects without any manual intervention. The method also identifies the (hopefully, few) visibility calculations which remain to be performed after the viewing position is specified. Also discussed is the development of still stronger solutions which could further reduce the number of these visibility calculations remaining at image generation time
keywords algorithms, hidden lines, hidden surfaces, computer graphics
series CADline
last changed 2003/06/02 13:58

_id 0868
authors Gero, John S. and Volfneuk, M.
year 1979
title Building Fuzzy CAD Systems
source 1979? pp. 74-79 : ill. includes bibliography
summary The paper introduces the need to include subjectivities in computer aided design systems. It commences with the differences between uncertainty, which has been used to model subjectivity, and imprecision. The former provides the basis of probability theory, whilst the latter the basis of fuzzy set theory. The thesis is that subjectivities introduce imprecision. It shows that subjectivities can be included in the description of the interactions between parts of the system. After presenting a brief introduction to fuzzy set theory the paper shows how a fuzzy CAD system can be built. An example is presented which demonstrates the approach
keywords CAD, fuzzy logic
series CADline
email
last changed 2003/06/02 13:58

_id c3b5
authors Hinds, John K. and Kuan, L.P.
year 1979
title Sculptured Surface Technology as a Unified Approach to Geometric Definition
source CASA - The Computer and Automated System Association of SME. 23 p. : ill Dearborn: SME, 1979. MS79-146. includes bibliography.
summary The purpose of this paper is to describe a comprehensive approach to representing and machining complex surface shapes in an APT programming system. The APT (Automatically Programmed Tools) user language was extended to permit the definition of a hierarchy of curves and surfaces. Much of the logic has been implemented using matrix canonical forms which are closed under the full family of projective transformations, permitting family of parts storage and retrieval and part compensation. The area of numerical control machining was addressed, but the solutions for tool positioning were only partially successful due to the complexity of the algorithmic problem. This paper first outlines some of the mathematical methods adopted and then illustrates how these have been implemented with an APT part programming example
keywords curved surfaces, representation, geometric modeling, mechanical engineering, CAM
series CADline
last changed 2003/06/02 13:58

_id a4cc
authors Hunter, Gregory M. and Steiglitz, Kenneth
year 1979
title Operations on Images Using Quad Trees
source IEEE Transactions on Pattern Analysis and Machine Intelligence. April, 1979. vol. PAMI-1: pp. 145-153 : ill. includes bibliography
summary A quad tree for representing a picture is a tree in which successively deeper levels represent successively finer subdivisions of picture area. An algorithm is given for superposing N quad trees in time proportional to the total number of nodes in the trees. Warnock-type algorithms are then presented for building the quad tree for the picture of the boundary of a polygon, and for coloring the interior of such a polygon. These algorithms take O(v + p + q) time, where v is the number of polygon vertices, p is the polygon perimeter, and q is a resolution parameter. When the resolution q is fixed, these algorithms are asymptotically optimal
keywords GIS, mapping, CAD, layout, pattern recognition, polyhedra, quadtree, planning, image processing, computer graphics
series CADline
last changed 2003/06/02 13:58

_id 2ccd
authors Kalisperis, Loukas N.
year 1994
title 3D Visualization in Design Education
doi https://doi.org/10.52842/conf.acadia.1994.177
source Reconnecting [ACADIA Conference Proceedings / ISBN 1-880250-03-9] Washington University (Saint Louis / USA) 1994, pp. 177-184
summary It has been said that "The beginning of architecture is empty space." (Mitchell 1990) This statement typifies a design education philosophy in which the concepts of space and form are separated and defined respectively as the negative and positive of the physical world, a world where solid objects exist and void-the mere absence of substance-is a surrounding atmospheric emptiness. Since the beginning of the nineteenth century, however, there has been an alternative concept of space as a continuum: that there is a continuously modified surface between the pressures of form and space in which the shape of the space in our lungs is directly connected to the shape of the space within which we exist. (Porter 1979). The nature of the task of representing architecture alters to reflect the state of architectural understanding at each period of time. The construction of architectural space and form represents a fundamental achievement of humans in their environment and has always involved effort and materials requiring careful planning, preparation, and forethought. In architecture there is a necessary conversion to that which is habitable, experiential, and functional from an abstraction in an entirely different medium. It is often an imperfect procedure that centers on the translation rather than the actual design. Design of the built environment is an art of distinctions within the continuum of space, for example: between solid and void, interior and exterior, light and dark, or warm and cold. It is concerned with the physical organization and articulation of space. The amount and shape of the void contained and generated by the building create the fabric and substance of the built environment. Architecture as a design discipline, therefore, can be considered as a creative expression of the coexistence of form and space on a human scale. As Frank Ching writes in Architecture: Form, Space, and Order, "These elements of form and space are the critical means of architecture. While the utilitarian concerns of function and use can be relatively short lived, and symbolic interpretations can vary from age to age, these primary elements of form and space comprise timeless and fundamental vocabulary of the architectural designer." (1979)

series ACADIA
email
last changed 2022/06/07 07:52

_id c6a9
authors Kay, Douglas Scott and Greenberg, Donald P.
year 1979
title Transparency for Computer Synthesized Images
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 158-164 : ill. (some col.). includes bibliography
summary Simple transparency algorithms which assume a linear transparency over an entire surface are the type most often employed to produce computer synthesized images of transparent objects with curved surfaces. Although most of the images created with these algorithms do give the impression of transparency, they usually do not look realistic. One of the most serious problems is that the intensity of the light that is transmitted through the objects is generally not proportional to the amount of material through which it must pass. Another problem is that the image seen behind the objects is not distorted as would naturally occur when the light is refracted as it passes through a material of different density. Use of a non-linear transparency algorithm can provide a great improvement in the realism of an image at a small additional cost. Making the transparency proportional to the normal to the surface causes it to decrease towards the edges of the surface where the path of the light through the object is longer. The exact simulation of refraction, however, requires that each sight ray be individually traced from the observer, through the picture plane and through each transparent object until an opaque surface is intersected. Since the direction of the ray would change as each material of differing optical density was entered, the hidden surface calculations required would be very time consuming. However, if a few assumptions are made about the geometry of each object and about the conditions under which they are viewed, a much simpler algorithm can be used to approximate the refractive effect. This method proceeds in a back-to-front order, mapping the current background image onto the next surface, until all surfaces have been considered
keywords computer graphics, shading, transformation, display, visualization, algorithms, realism
series CADline
last changed 2003/06/02 13:58

_id 8023
authors Lang, M.S., Cohen, R.L. and Eschenberg, K.E. (et al)
year 1979
title Implementation of An Interactive Computer Graphics Environment at NASA/JSC
source SIGGRAPH '79 Conference Proceedings. August, 1979. vol. 13 ; no. 2: pp. 246-252 : ill. includes bibliography
summary The implementation of visually-oriented software for graphics support on the high-performance computer graphics hardware at NASA's Johnson Space Center is the latest step in the evolution of an interactive computer applications technology being developed by the Computer Graphics Group at The Applied Research Laboratory of Penn State University. This technology is designed to aid the typical scientist or engineer in learning and using computer graphics productively, including writing his own programs and interfacing to software specialists who will write and maintain his programs. Key aspects of the current development include the creation and incorporation of a visually-oriented learning package for graphics geometric perception and graphics programming, as well as a sophisticated control environment which aides the user in obtaining a quick understanding of and access to the system. Preliminary results indicate that this software support can substantially reduce the start-up time for a novice graphics user with some background in Fortran
keywords computer graphics, user interface, software, learning, programming, control, education
series CADline
last changed 2003/06/02 13:58

_id ddss2006-hb-187
id DDSS2006-HB-187
authors Lidia Diappi and Paola Bolchi
year 2006
title Gentrification Waves in the Inner-City of Milan - A multi agent / cellular automata model based on Smith's Rent Gap theory
source Van Leeuwen, J.P. and H.J.P. Timmermans (eds.) 2006, Innovations in Design & Decision Support Systems in Architecture and Urban Planning, Dordrecht: Springer, ISBN-10: 1-4020-5059-3, ISBN-13: 978-1-4020-5059-6, p. 187-201
summary The aim of this paper is to investigate the gentrification process by applying an urban spatial model of gentrification, based on Smith's (1979; 1987; 1996) Rent Gap theory. The rich sociological literature on the topic mainly assumes gentrification to be a cultural phenomenon, namely the result of a demand pressure of the suburban middle and upper class, willing to return to the city (Ley, 1980; Lipton, 1977, May, 1996). Little attempt has been made to investigate and build a sound economic explanation on the causes of the process. The Rent Gap theory (RGT) of Neil Smith still represents an important contribution in this direction. At the heart of Smith's argument there is the assumption that gentrification takes place because capitals return to the inner city, creating opportunities for residential relocation and profit. This paper illustrates a dynamic model of Smith's theory through a multi-agent/ cellular automata system approach (Batty, 2005) developed on a Netlogo platform. A set of behavioural rules for each agent involved (homeowner, landlord, tenant and developer, and the passive 'dwelling' agent with their rent and level of decay) are formalised. The simulations show the surge of neighbouring degradation or renovation and population turn over, starting with different initial states of decay and estate rent values. Consistent with a Self Organized Criticality approach, the model shows that non linear interactions at local level may produce different configurations of the system at macro level. This paper represents a further development of a previous version of the model (Diappi, Bolchi, 2005). The model proposed here includes some more realistic factors inspired by the features of housing market dynamics in the city of Milan. It includes the shape of the potential rent according to city form and functions, the subdivision in areal submarkets according to the current rents, and their maintenance levels. The model has a more realistic visualisation of the city and its form, and is able to show the different dynamics of the emergent neighbourhoods in the last ten years in Milan.
keywords Multi agent systems, Housing market, Gentrification, Emergent systems
series DDSS
last changed 2006/08/29 12:55

_id 69b3
authors Markelin, Antero
year 1993
title Efficiency of Model Endoscopic Simulation - An Experimental Research at the University of Stuttgart
source Endoscopy as a Tool in Architecture [Proceedings of the 1st European Architectural Endoscopy Association Conference / ISBN 951-722-069-3] Tampere (Finland), 25-28 August 1993, pp. 31-34
summary At the Institute of Urban Planning at the University of Stuttgart early experiments were made with the help of endoscopes in the late 1970’s. The intention was to find new instruments to visualize urban design projects. The first experiment included the use of a 16 mm film of a 1:170 scale model of the market place at Karlsruhe, including design alternatives (with trees, without trees etc). The film was shown to the Karlsruhe authorities, who had to make the decision about the alternatives. It was said, that the film gave a great help for the decision-making and a design proposition had never before been presented in such understandable way. In 1975-77, with the support of the Deutsche Forschungsgemeinschaft (German Research Foundation) an investigation was carried out into existing endoscopic simulation facilities, such as those in Wageningen, Lund and Berkeley. The resulting publication was mainly concerned with technical installations and their applications. However a key question remained: ”Can reality be simulated with endoscopy?” In 1979-82, in order to answer that question, at the Institute was carried out the most extensive research of the time, into the validity of endoscopic simulation. Of special importance was the inclusion of social scientists and psychologists from the University of Heidelberg and Mannheim. A report was produced in 1983. The research was concerned with the theory of model simulation, its ways of use and its users, and then the establishment of requirements for effective model simulation. For the main research work with models or simulation films, psychological tests were developed which enabled a tested person to give accurate responses or evidence without getting involved in alien technical terminology. It was also thought that the use of semantic differentials would make the work imprecise or arbitrary.

keywords Architectural Endoscopy
series EAEA
more http://info.tuwien.ac.at/eaea/
last changed 2005/09/09 10:43

For more results click below:

this is page 0show page 1HOMELOGIN (you are user _anon_171149 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002