CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 36

_id 20a5
authors Kieburtz, Richard B.
year 1977
title Structured Programming and Problem- Solving with PASCAL
source xiii, 348 p. : ill. Englewood cliffs, New Jersey: Prentice-Hall, Inc., 1977. includes index
summary An introduction emphasizing the problem-solving approach to computing, progressing from the development of a systematic and disciplined approach to the discovery of algorithms. Includes examples and exercises
keywords PASCAL, programming, languages, problem solving, education
series CADline
last changed 2003/06/02 13:58

_id ecaade2009_177
id ecaade2009_177
authors Göttig, Roland; Braunes, Jörg
year 2009
title Building Survey in Combination with Building Information Modelling for the Architectural Planning Process
source Computation: The New Realm of Architectural Design [27th eCAADe Conference Proceedings / ISBN 978-0-9541183-8-9] Istanbul (Turkey) 16-19 September 2009, pp. 69-74
doi https://doi.org/10.52842/conf.ecaade.2009.069
wos WOS:000334282200007
summary The architectural planning process is influenced by social, cultural and technical aspects (Alexander, 1977). When focussing on computer based planning for retrofitting or modification of buildings it becomes clear that many different data formats are used depending on a great variety of planning methods. Moreover, if building information models are utilized they still lack some essential criteria. It is rarely possible to attach individual data from survey systems. This paper will show both a way to add data from building survey systems as an example for special data attachment on IFC files and how to utilize content management systems for IFC files, deviated plans, lists of building components, and other data necessary in a planning process.
keywords Planning process, building information modeling, IFC, building survey systems, content management systems
series eCAADe
email
last changed 2022/06/07 07:50

_id sigradi2013_41
id sigradi2013_41
authors Luhan, Gregory A.; Robert Gregory
year 2013
title Across Disciplines: Triggering Frame Awareness in Design Education
source SIGraDi 2013 [Proceedings of the 17th Conference of the Iberoamerican Society of Digital Graphics - ISBN: 978-956-7051-86-1] Chile - Valparaíso 20 - 22 November 2013, pp. 619 - 623
summary Tacit knowledge is paradoxical: something we know yet don't know we know, knowledge we sense but can't articulate. In Polanyi’s definition of tacit knowledge, “we know more than we can say" (1966/2009; Scott, 1985; Gelwick, 1977). It's important to see that tacit knowledge is part of a sequence; mental structures, in awareness when first learned, eventually become tacit, operating thenceforth as unquestioned assumptions. These tacit structures pose a problem for professional education in disciplines that encourage creativity. This paper examines the design and re-design of an interdisciplinary course intended to help make these tacit structures visible, to trigger frame awareness.
keywords Tacit knowledge; Design thinking; Sustainability; Systems thinking; Frame reflection
series SIGRADI
email
last changed 2016/03/10 09:55

_id caadria2023_395
id caadria2023_395
authors Luo, Jiaxiang, Mastrokalou, Efthymia, Aldaboos, Sarah and Aldabous, Rahaf
year 2023
title Research on the Exploration of Sprayed Clay Material and Modeling System
source Immanuel Koh, Dagmar Reinhardt, Mohammed Makki, Mona Khakhar, Nic Bao (eds.), HUMAN-CENTRIC - Proceedings of the 28th CAADRIA Conference, Ahmedabad, 18-24 March 2023, pp. 231–240
doi https://doi.org/10.52842/conf.caadria.2023.2.231
summary As a traditional building material, clay has been used by humans for a long time. From early civilisations, to the modern dependence on new technologies, the craft of clay making is commonly linked with the use of moulds, handmade creations, ceramic extruders, etc. (Schmandt and Besserat, 1977). Clay in the form of bricks is one of the oldest building materials known (Fernandes et al, 2010). This research expands the possibilities offered by standardised bricks by testing types of clay, forms, shapes, porosity, and structural methods. The traditional way of working with clay relies on human craftsmanship and is based on the use of semi-solid clay (Fernandes et al., 2010). However, there is little research on the use of clay slurry. With the rise of 3D printing systems in recent years, research and development has been emerging on using clay as a 3D printing filament (Gürsoy, 2018). Researchers have discovered that in order for 3D-printed clay slurry to solidify quickly to support the weight of the added layers during printing, curing agents such as lime, coal ash, cement, etc. have to be added to the clay slurry. After adding these substances, clay is difficult to be reused and can have a negative effect on the environment (Chen et al., 2021). In this study, a unique method for manufacturing clay elements of intricate geometries is proposed with the help of an internal skeleton that can be continuously reused. The study introduces the process of applying clay on a special structure through spraying and showcases how this method creates various opportunities for customisation of production.
keywords Spray clay, Substructure, 3D printing, Modelling system, Reusable
series CAADRIA
email
last changed 2023/06/15 23:14

_id c7e9
authors Maver, T.W.
year 2002
title Predicting the Past, Remembering the Future
source SIGraDi 2002 - [Proceedings of the 6th Iberoamerican Congress of Digital Graphics] Caracas (Venezuela) 27-29 november 2002, pp. 2-3
summary Charlas Magistrales 2There never has been such an exciting moment in time in the extraordinary 30 year history of our subject area, as NOW,when the philosophical theoretical and practical issues of virtuality are taking centre stage.The PastThere have, of course, been other defining moments during these exciting 30 years:• the first algorithms for generating building layouts (circa 1965).• the first use of Computer graphics for building appraisal (circa 1966).• the first integrated package for building performance appraisal (circa 1972).• the first computer generated perspective drawings (circa 1973).• the first robust drafting systems (circa 1975).• the first dynamic energy models (circa 1982).• the first photorealistic colour imaging (circa 1986).• the first animations (circa 1988)• the first multimedia systems (circa 1995), and• the first convincing demonstrations of virtual reality (circa 1996).Whereas the CAAD community has been hugely inventive in the development of ICT applications to building design, it hasbeen woefully remiss in its attempts to evaluate the contribution of those developments to the quality of the built environmentor to the efficiency of the design process. In the absence of any real evidence, one can only conjecture regarding the realbenefits which fall, it is suggested, under the following headings:• Verisimilitude: The extraordinary quality of still and animated images of the formal qualities of the interiors and exteriorsof individual buildings and of whole neighborhoods must surely give great comfort to practitioners and their clients thatwhat is intended, formally, is what will be delivered, i.e. WYSIWYG - what you see is what you get.• Sustainability: The power of «first-principle» models of the dynamic energetic behaviour of buildings in response tochanging diurnal and seasonal conditions has the potential to save millions of dollars and dramatically to reduce thedamaging environmental pollution created by badly designed and managed buildings.• Productivity: CAD is now a multi-billion dollar business which offers design decision support systems which operate,effectively, across continents, time-zones, professions and companies.• Communication: Multi-media technology - cheap to deliver but high in value - is changing the way in which we canexplain and understand the past and, envisage and anticipate the future; virtual past and virtual future!MacromyopiaThe late John Lansdown offered the view, in his wonderfully prophetic way, that ...”the future will be just like the past, onlymore so...”So what can we expect the extraordinary trajectory of our subject area to be?To have any chance of being accurate we have to have an understanding of the phenomenon of macromyopia: thephenomenon exhibitted by society of greatly exaggerating the immediate short-term impact of new technologies (particularlythe information technologies) but, more importantly, seriously underestimating their sustained long-term impacts - socially,economically and intellectually . Examples of flawed predictions regarding the the future application of information technologiesinclude:• The British Government in 1880 declined to support the idea of a national telephonic system, backed by the argumentthat there were sufficient small boys in the countryside to run with messages.• Alexander Bell was modest enough to say that: «I am not boasting or exaggerating but I believe, one day, there will bea telephone in every American city».• Tom Watson, in 1943 said: «I think there is a world market for about 5 computers».• In 1977, Ken Olssop of Digital said: «There is no reason for any individuals to have a computer in their home».The FutureJust as the ascent of woman/man-kind can be attributed to her/his capacity to discover amplifiers of the modest humancapability, so we shall discover how best to exploit our most important amplifier - that of the intellect. The more we know themore we can figure; the more we can figure the more we understand; the more we understand the more we can appraise;the more we can appraise the more we can decide; the more we can decide the more we can act; the more we can act themore we can shape; and the more we can shape, the better the chance that we can leave for future generations a trulysustainable built environment which is fit-for-purpose, cost-beneficial, environmentally friendly and culturally significactCentral to this aspiration will be our understanding of the relationship between real and virtual worlds and how to moveeffortlessly between them. We need to be able to design, from within the virtual world, environments which may be real ormay remain virtual or, perhaps, be part real and part virtual.What is certain is that the next 30 years will be every bit as exciting and challenging as the first 30 years.
series SIGRADI
email
last changed 2016/03/10 09:55

_id ecaade2007_042
id ecaade2007_042
authors Ozel, Filiz
year 2007
title Pattern Language and Embedded Knowledge in Building Information Modeling
source Predicting the Future [25th eCAADe Conference Proceedings / ISBN 978-0-9541183-6-5] Frankfurt am Main (Germany) 26-29 September 2007, pp. 457-464
doi https://doi.org/10.52842/conf.ecaade.2007.457
summary When Christopher Alexander (1977), trained both as a mathematician and an architect, published his seminal work “The Pattern Language” in the 1970’s and introduced the concept of “pattern language”, computers were still in their infancy, CAD did not exist as we know it today, and computer information modeling was not even in the radar screen of researchers. Design communication simply meant manual drafting. With the concept of ‘pattern language” (http://www.patternlanguage.com/), Alexander proposed a systematic method for dealing with complexity, which proved itself to be more relevant than ever in the digital age. The concept is often cited by computer scientists as a precursor to object oriented modeling. This study explores the potential of “pattern language” for structuring building information and design knowledge within the framework of the recent developments in building information modeling (BIM). In this article, comparisons to the approach taken by the software engineering industry who embraced the idea of “patterns” as a systematic way to software development are also made. While Alexander’s pattern language proposes a method with which the designer can incorporate his/her experiences and design vision systematically into the process of designing, software industry’s approach to patterns describes a method for providing problem and solution patterns (i.e. prototypes) that can be used repeatedly during software development. There is obviously a significant difference between the original intent of the “pattern language” and the way it was later used in other fields including software engineering and business solutions. At the cross section of architectural design and software engineering, Building Information Modeling (BIM) software can benefit from carefully incorporating a combination of these two approaches into its structure as patterns.
keywords Building information modeling, Christopher Alexander, pattern language, software development
series eCAADe
email
last changed 2022/06/07 08:00

_id 80cb
authors Winston, Patrick Henry
year 1977
title Artificial Intelligence
source xvi, 444 p. : ill. Reading, Mass.: Addison Wesley Pub. Co., 1977. includes bibliography and index. -- ( Addison-Wesley Series in Computer Science)
summary An introduction to key ideas in artificial intelligence and the use of LISP and other derived languages
keywords AI, LISP
series CADline
last changed 1999/02/12 15:10

_id ddssar0206
id ddssar0206
authors Bax, M.F.Th. and Trum, H.M.G.J.
year 2002
title Faculties of Architecture
source Timmermans, Harry (Ed.), Sixth Design and Decision Support Systems in Architecture and Urban Planning - Part one: Architecture Proceedings Avegoor, the Netherlands), 2002
summary In order to be inscribed in the European Architect’s register the study program leading to the diploma ‘Architect’ has to meet the criteria of the EC Architect’s Directive (1985). The criteria are enumerated in 11 principles of Article 3 of the Directive. The Advisory Committee, established by the European Council got the task to examine such diplomas in the case some doubts are raised by other Member States. To carry out this task a matrix was designed, as an independent interpreting framework that mediates between the principles of Article 3 and the actual study program of a faculty. Such a tool was needed because of inconsistencies in the list of principles, differences between linguistic versions ofthe Directive, and quantification problems with time, devoted to the principles in the study programs. The core of the matrix, its headings, is a categorisation of the principles on a higher level of abstractionin the form of a taxonomy of domains and corresponding concepts. Filling in the matrix means that each study element of the study programs is analysed according to their content in terms of domains; thesummation of study time devoted to the various domains results in a so-called ‘profile of a faculty’. Judgement of that profile takes place by committee of peers. The domains of the taxonomy are intrinsically the same as the concepts and categories, needed for the description of an architectural design object: the faculties of architecture. This correspondence relates the taxonomy to the field of design theory and philosophy. The taxonomy is an application of Domain theory. This theory,developed by the authors since 1977, takes as a view that the architectural object only can be described fully as an integration of all types of domains. The theory supports the idea of a participatory andinterdisciplinary approach to design, which proved to be awarding both from a scientific and a social point of view. All types of domains have in common that they are measured in three dimensions: form, function and process, connecting the material aspects of the object with its social and proceduralaspects. In the taxonomy the function dimension is emphasised. It will be argued in the paper that the taxonomy is a categorisation following the pragmatistic philosophy of Charles Sanders Peirce. It will bedemonstrated as well that the taxonomy is easy to handle by giving examples of its application in various countries in the last 5 years. The taxonomy proved to be an adequate tool for judgement ofstudy programs and their subsequent improvement, as constituted by the faculties of a Faculty of Architecture. The matrix is described as the result of theoretical reflection and practical application of a matrix, already in use since 1995. The major improvement of the matrix is its direct connection with Peirce’s universal categories and the self-explanatory character of its structure. The connection with Peirce’s categories gave the matrix a more universal character, which enables application in other fieldswhere the term ‘architecture’ is used as a metaphor for artefacts.
series DDSS
last changed 2003/11/21 15:16

_id 4489
authors Blinn, J.F.
year 1977
title Models of light reflection for computer synthesised pictures
source Computer Graphics, 11 2, 192-198
summary Bui-Tuong Phong published his illumination model in 1973, in the paper titled "Illumination for Computer-Generated Images". Phong's model is a local illumination model, which means only direct reflections are taken into account. Light that bounces off more than one surface before reaching the eye is not accounted for. While this may not be very realistic, it allows the lighting to be computed efficiently. To properly handle indirect lighting, a global illumination method such as radiosity is required, which is much more expensive. In addition to Phong's basic lighting equation, we will look at a variation invented by Jim Blinn. Blinn changed the way specular is calculated, making the computations slightly cheaper. Blinn published his approach in his paper "Models of Light Reflection for Computer Synthesised Pictures" in 1977.
series journal paper
last changed 2003/04/23 15:14

_id 2168
authors Bobrow, Daniel G. and Winograd, Terry
year 1977
title An Overview of KRL, a Knowledge Representation Language
source Cognitive Science. 1977. vol. 1: pp. 3-46. includes bibliography
summary This paper describes KRL, a Knowledge Representation Language designed for use in understander systems. It outlines both the general concepts which underlie the research and the details of KRL-O, an experimental implementation of some of these concepts. KRL is an attempt to integrate procedural knowledge with a broad base of declarative forms. These forms provide a variety of ways to express the logical structure of the knowledge, in order to give flexibility in associating procedures (for memory and reasoning) with specific pieces of knowledge, and to control the relative accessibility of different facts and descriptions. The formalism for declarative knowledge is based on structured conceptual objects with associated descriptions. These objects form a network of memory units with several different sorts of linkages, each having well-specified implications for the retrieval process. Procedures can be associated directly with the internal structure of a conceptual object. This procedural attachment allows the steps for a particular operation to be determined by characteristics of the specific entities involved. The control structure of KRL is based on the belief that the next generation of intelligent programs will integrate data-directed and goal-directed processing by using multiprocessing. It provides for a priority-ordered multiprocess agenda with explicit (user-provided) strategies for scheduling and resource allocation. It provides procedure directories which operate along with process frameworks to allow procedural parametrization of the fundamental system processes for building, comparing, and retrieving memory structures. Future development of KRL will include integrating procedure definition with the descriptive formalism
keywords knowledge, representation, languages, AI
series CADline
last changed 2003/06/02 10:24

_id aef9
id aef9
authors Brown, A., Knight, M. and Berridge, P. (Eds.)
year 1999
title Architectural Computing from Turing to 2000 [Conference Proceedings]
source eCAADe Conference Proceedings / ISBN 0-9523687-5-7 / Liverpool (UK) 15-17 September 1999, 773 p.
doi https://doi.org/10.52842/conf.ecaade.1999
summary The core theme of this book is the idea of looking forward to where research and development in Computer Aided Architectural Design might be heading. The contention is that we can do so most effectively by using the developments that have taken place over the past three or four decades in Computing and Architectural Computing as our reference point; the past informing the future. The genesis of this theme is the fact that a new millennium is about to arrive. If we are ruthlessly objective the year 2000 holds no more significance than any other year; perhaps we should, instead, be preparing for the year 2048 (2k). In fact, whatever the justification, it is now timely to review where we stand in terms of the development of Architectural Computing. This book aims to do that. It is salutary to look back at what writers and researchers have said in the past about where they thought that the developments in computing were taking us. One of the common themes picked up in the sections of this book is the developments that have been spawned by the global linkup that the worldwide web offers us. In the past decade the scale and application of this new medium of communication has grown at a remarkable rate. There are few technological developments that have become so ubiquitous, so quickly. As a consequence there are particular sections in this book on Communication and the Virtual Design Studio which reflect the prominence of this new area, but examples of its application are scattered throughout the book. In 'Computer-Aided Architectural Design' (1977), Bill Mitchell did suggest that computer network accessibility from expensive centralised locations to affordable common, decentralised computing facilities would become more commonplace. But most pundits have been taken by surprise by just how powerful the explosive cocktail of networks, email and hypertext has proven to be. Each of the ingredients is interesting in its own right but together they have presented us with genuinely new ways of working. Perhaps, with foresight we can see what the next new explosive cocktail might be.
series eCAADe
email
more http://www.ecaade.org
last changed 2022/06/07 07:49

_id sigradi2009_774
id sigradi2009_774
authors de Souza, Raphael Argento; André Soares Monat
year 2009
title Visualização da Informação em meio telejornalístico: Uma abordagem sob a ótica do design [Information Visualization in the news television: An approach under the design sight]
source SIGraDi 2009 - Proceedings of the 13th Congress of the Iberoamerican Society of Digital Graphics, Sao Paulo, Brazil, November 16-18, 2009
summary This article proposes a classification, under the Visualization Information point of view, of infographics broadcasted in the brazilian news television. To achieve this purpose these so called motion graphics were analised under the basis formed by three main authors: Tufte (1997), Bertin (1977) and Spence (2007), whose theories are in this article compared to the digital means of the motion graphics. With these theoretical foundation and the analisys of two hundred motion graphics broadcasted in the brazilian news television, we achieved a classification which covers every type of these motion graphics, hoping it becomes a basis for the study of these projects.
keywords Design; information visualization; television infographics, motion graphics; information design
series SIGRADI
email
last changed 2016/03/10 09:50

_id sigradi2006_e028c
id sigradi2006_e028c
authors Griffith, Kenfield; Sass, Larry and Michaud, Dennis
year 2006
title A strategy for complex-curved building design:Design structure with Bi-lateral contouring as integrally connected ribs
source SIGraDi 2006 - [Proceedings of the 10th Iberoamerican Congress of Digital Graphics] Santiago de Chile - Chile 21-23 November 2006, pp. 465-469
summary Shapes in designs created by architects such as Gehry Partners (Shelden, 2002), Foster and Partners, and Kohn Peterson and Fox rely on computational processes for rationalizing complex geometry for building construction. Rationalization is the reduction of a complete geometric shape into discrete components. Unfortunately, for many architects the rationalization is limited reducing solid models to surfaces or data on spread sheets for contractors to follow. Rationalized models produced by the firms listed above do not offer strategies for construction or digital fabrication. For the physical production of CAD description an alternative to the rationalized description is needed. This paper examines the coupling of digital rationalization and digital fabrication with physical mockups (Rich, 1989). Our aim is to explore complex relationships found in early and mid stage design phases when digital fabrication is used to produce design outcomes. Results of our investigation will aid architects and engineers in addressing the complications found in the translation of design models embedded with precision to constructible geometries. We present an algorithmically based approach to design rationalization that supports physical production as well as surface production of desktop models. Our approach is an alternative to conventional rapid prototyping that builds objects by assembly of laterally sliced contours from a solid model. We explored an improved product description for rapid manufacture as bilateral contouring for structure and panelling for strength (Kolarevic, 2003). Infrastructure typically found within aerospace, automotive, and shipbuilding industries, bilateral contouring is an organized matrix of horizontal and vertical interlocking ribs evenly distributed along a surface. These structures are monocoque and semi-monocoque assemblies composed of structural ribs and skinning attached by rivets and adhesives. Alternative, bi-lateral contouring discussed is an interlocking matrix of plywood strips having integral joinery for assembly. Unlike traditional methods of building representations through malleable materials for creating tangible objects (Friedman, 2002), this approach constructs with the implication for building life-size solutions. Three algorithms are presented as examples of rationalized design production with physical results. The first algorithm [Figure 1] deconstructs an initial 2D curved form into ribbed slices to be assembled through integral connections constructed as part of the rib solution. The second algorithm [Figure 2] deconstructs curved forms of greater complexity. The algorithm walks along the surface extracting surface information along horizontal and vertical axes saving surface information resulting in a ribbed structure of slight double curvature. The final algorithm [Figure 3] is expressed as plug-in software for Rhino that deconstructs a design to components for assembly as rib structures. The plug-in also translates geometries to a flatten position for 2D fabrication. The software demonstrates the full scope of the research exploration. Studies published by Dodgson argued that innovation technology (IvT) (Dodgson, Gann, Salter, 2004) helped in solving projects like the Guggenheim in Bilbao, the leaning Tower of Pisa in Italy, and the Millennium Bridge in London. Similarly, the method discussed in this paper will aid in solving physical production problems with complex building forms. References Bentley, P.J. (Ed.). Evolutionary Design by Computers. Morgan Kaufman Publishers Inc. San Francisco, CA, 1-73 Celani, G, (2004) “From simple to complex: using AutoCAD to build generative design systems” in: L. Caldas and J. Duarte (org.) Implementations issues in generative design systems. First Intl. Conference on Design Computing and Cognition, July 2004 Dodgson M, Gann D.M., Salter A, (2004), “Impact of Innovation Technology on Engineering Problem Solving: Lessons from High Profile Public Projects,” Industrial Dynamics, Innovation and Development, 2004 Dristas, (2004) “Design Operators.” Thesis. Massachusetts Institute of Technology, Cambridge, MA, 2004 Friedman, M, (2002), Gehry Talks: Architecture + Practice, Universe Publishing, New York, NY, 2002 Kolarevic, B, (2003), Architecture in the Digital Age: Design and Manufacturing, Spon Press, London, UK, 2003 Opas J, Bochnick H, Tuomi J, (1994), “Manufacturability Analysis as a Part of CAD/CAM Integration”, Intelligent Systems in Design and Manufacturing, 261-292 Rudolph S, Alber R, (2002), “An Evolutionary Approach to the Inverse Problem in Rule-Based Design Representations”, Artificial Intelligence in Design ’02, 329-350 Rich M, (1989), Digital Mockup, American Institute of Aeronautics and Astronautics, Reston, VA, 1989 Schön, D., The Reflective Practitioner: How Professional Think in Action. Basic Books. 1983 Shelden, D, (2003), “Digital Surface Representation and the Constructability of Gehry’s Architecture.” Diss. Massachusetts Institute of Technology, Cambridge, MA, 2003 Smithers T, Conkie A, Doheny J, Logan B, Millington K, (1989), “Design as Intelligent Behaviour: An AI in Design Thesis Programme”, Artificial Intelligence in Design, 293-334 Smithers T, (2002), “Synthesis in Designing”, Artificial Intelligence in Design ’02, 3-24 Stiny, G, (1977), “Ice-ray: a note on the generation of Chinese lattice designs” Environmental and Planning B, volume 4, pp. 89-98
keywords Digital fabrication; bilateral contouring; integral connection; complex-curve
series SIGRADI
email
last changed 2016/03/10 09:52

_id 76ce
authors Grimson, W.
year 1985
title Computational Experiments with a Feature Based Stereo Algorithm
source IEEE Trans. Pattern Anal. Machine Intell., Vol. PAMI-7, No. 1
summary Computational models of the human stereo system' can provide insight into general information processing constraints that apply to any stereo system, either artificial or biological. In 1977, Marr and Poggio proposed one such computational model, that was characterized as matching certain feature points in difference-of-Gaussian filtered images, and using the information obtained by matching coarser resolution representations to restrict the search'space for matching finer resolution representations. An implementation of the algorithm and'its testing on a range of images was reported in 1980. Since then a number of psychophysical experiments have suggested possible refinements to the model and modifications to the algorithm. As well, recent computational experiments applying the algorithm to a variety of natural images, especially aerial photographs, have led to a number of modifications. In this article, we present a version of the Marr-Poggio-Gfimson algorithm that embodies these modifications and illustrate its performance on a series of natural images.
series journal paper
last changed 2003/04/23 15:14

_id a4b7
authors Lee, D. T. and Preparata, Franco P.
year 1977
title Location of a Point in a Planar Subdivision and its Applications
source SIAM Journal of Computing. September, 1977. vol. 6: pp. 594-606 : ill. includes bibliography
summary Given a subdivision of the plane induced by a planar graph with n vertices, in this paper the problem of identifying which region of the subdivision contains a given test points is considered. A search algorithm, called point-location algorithm, which operates on a suitably preprocessed data structure is presented. The search runs in time at most O((log n)2), while the preprocessing task runs in time at most O(n log n) and requires O(n) storage. The methods are quite general, since an arbitrary subdivision can be transformed in time at most O(n log n) into one to which the preprocessing procedure is applicable. This solution of the point location problem yields interesting and efficient solutions of other geometric problems, such as spatial convex inclusion and inclusion in an arbitrary polygon
keywords computational geometry, algorithms, analysis, graphs, point inclusion
series CADline
last changed 2003/06/02 13:58

_id ddss2006-hb-187
id DDSS2006-HB-187
authors Lidia Diappi and Paola Bolchi
year 2006
title Gentrification Waves in the Inner-City of Milan - A multi agent / cellular automata model based on Smith's Rent Gap theory
source Van Leeuwen, J.P. and H.J.P. Timmermans (eds.) 2006, Innovations in Design & Decision Support Systems in Architecture and Urban Planning, Dordrecht: Springer, ISBN-10: 1-4020-5059-3, ISBN-13: 978-1-4020-5059-6, p. 187-201
summary The aim of this paper is to investigate the gentrification process by applying an urban spatial model of gentrification, based on Smith's (1979; 1987; 1996) Rent Gap theory. The rich sociological literature on the topic mainly assumes gentrification to be a cultural phenomenon, namely the result of a demand pressure of the suburban middle and upper class, willing to return to the city (Ley, 1980; Lipton, 1977, May, 1996). Little attempt has been made to investigate and build a sound economic explanation on the causes of the process. The Rent Gap theory (RGT) of Neil Smith still represents an important contribution in this direction. At the heart of Smith's argument there is the assumption that gentrification takes place because capitals return to the inner city, creating opportunities for residential relocation and profit. This paper illustrates a dynamic model of Smith's theory through a multi-agent/ cellular automata system approach (Batty, 2005) developed on a Netlogo platform. A set of behavioural rules for each agent involved (homeowner, landlord, tenant and developer, and the passive 'dwelling' agent with their rent and level of decay) are formalised. The simulations show the surge of neighbouring degradation or renovation and population turn over, starting with different initial states of decay and estate rent values. Consistent with a Self Organized Criticality approach, the model shows that non linear interactions at local level may produce different configurations of the system at macro level. This paper represents a further development of a previous version of the model (Diappi, Bolchi, 2005). The model proposed here includes some more realistic factors inspired by the features of housing market dynamics in the city of Milan. It includes the shape of the potential rent according to city form and functions, the subdivision in areal submarkets according to the current rents, and their maintenance levels. The model has a more realistic visualisation of the city and its form, and is able to show the different dynamics of the emergent neighbourhoods in the last ten years in Milan.
keywords Multi agent systems, Housing market, Gentrification, Emergent systems
series DDSS
last changed 2006/08/29 12:55

_id 0d3a
authors Mitchell, William J.
year 1977
title Computer-Aided Architectural Design
source 573 p. : ill. New York: Mason Charter Pub. Inc., 1977. includes bibliography: p. 515-559 and index
summary This book is intended to provide a comprehensive introduction to the fundamentals of computer-aided architectural design for the students of architecture, the architect in practice, and the computer professional who is interested in learning about this application area
keywords CAD, architecture, education
series CADline
email
last changed 2003/06/02 13:58

_id 3105
authors Novak, T.P., Hoffman, D.L., and Yung, Y.-F.
year 1996
title Modeling the structure of the flow experience
source INFORMS Marketing Science and the Internet Mini-Conference, MIT
summary The flow construct (Csikszentmihalyi 1977) has recently been proposed by Hoffman and Novak (1996) as essential to understanding consumer navigation behavior in online environments such as the World Wide Web. Previous researchers (e.g. Csikszentmihalyi 1990; Ghani, Supnick and Rooney 1991; Trevino and Webster 1992; Webster, Trevino and Ryan 1993) have noted that flow is a useful construct for describing more general human-computer interactions. Hoffman and Novak define flow as the state occurring during network navigation which is: 1) characterized by a seamless sequence of responses facilitated by machine interactivity, 2) intrinsically enjoyable, 3) accompanied by a loss of self-consciousness, and 4) selfreinforcing." To experience flow while engaged in an activity, consumers must perceive a balance between their skills and the challenges of the activity, and both their skills and challenges must be above a critical threshold. Hoffman and Novak (1996) propose that flow has a number of positive consequences from a marketing perspective, including increased consumer learning, exploratory behavior, and positive affect."
series other
last changed 2003/04/23 15:50

_id 19a8
authors Schultz, Volkher
year 1999
title A DIDACTIC CONCEPT FOR TRAINING ARCHITECTS AND INTERIOR DECORATORS
source Full-scale Modeling and the Simulation of Light [Proceedings of the 7th European Full-scale Modeling Association Conference / ISBN 3-85437-167-5] Florence (Italy) 18-20 February 1999, pp. 29-39
summary The 20th year anniversary of the Lichtlabor is used as an opportunity to look back. Two decades in a period of rapid technological development is a long time, during which furnishings pass or fail their trial period. The organizational structure of the Lichtlabor which includes lay-out and appliances has continually expanded since 1977 although the theoretical approach has not changed. Even the ideational structure of the Lichtlabor, without which an organizational structure would be worthless, has proved to be workable and effective as a didactic concept. This concept is based on the interdisciplinary midpoint between a technically (basic) understanding of light – a combination of abstract knowledge and experience gained – and its design-related application.
keywords Lighting Laboratory, Interior Design, Didactic Concept, Visual Comfort, Architectural Space, Model Simulation, Real Environments
series other
type normal paper
more http://info.tuwien.ac.at/efa
last changed 2004/05/04 11:28

_id sigradi2020_128
id sigradi2020_128
authors Sousa, Megg; Mônaco, Denise; Martínez, Andressa; Souza, Douglas
year 2020
title The operationalization of "A Pattern Language" by using network analysis tools
source SIGraDi 2020 [Proceedings of the 24th Conference of the Iberoamerican Society of Digital Graphics - ISSN: 2318-6968] Online Conference 18 - 20 November 2020, pp. 128-136
summary One of the most significant public space studies, "A Pattern Language", published in 1977, is until today's background for some contemporary investigations. The aim of this paper is to propose an operationalization of the patterns' network of the book into a network analysis tool. The methodology is based on a new classification of patterns, in addition to what is initially presented in the book: "context patterns" (evidencing pre-existing conditions and potentialities) and "design patterns" (considering possibilities limited by the stakeholder at that location). The digital operationalization can enhance the analytical and predictive character of the work.
keywords Pattern language, Network analysis tool, Christopher Alexander, Public spaces
series SIGraDi
email
last changed 2021/07/16 11:48

For more results click below:

this is page 0show page 1HOMELOGIN (you are user _anon_911926 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002