CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 622

_id 39cb
authors Kelleners, Richard H.M.C.
year 1999
title Constraints in object-oriented graphics
source Eindhoven University of Technology
summary In the area of interactive computer graphics, two important approaches to deal with the complexity of designing and implementing graphics systems are object-oriented programming and constraint-based programming. From literature, it appears that combination of these two has clear advantages but has also proven to be difficult. One of the main problems is that constraint programming infringes the information hiding principle of object-oriented programming. The goal of the research project is to combine these two approaches to benefit from the strengths of both. Two research groups at the Eindhoven University of Technology investigate the use of constraints on graphics objects. At the Architecture department, constraints are applied in a virtual reality design environment. At the Computer Science department, constraints aid in modeling 3D animations. For these two groups, a constraint system for 3D graphical objects was developed. A conceptual model, called CODE (Constraints on Objects via Data flows and Events), is presented that enables integration of constraints and objects by separating the object world from the constraint world. In the design of this model, the main aspect being considered is that the information hiding principle among objects may not be violated. Constraint solvers, however, should have direct access to an object’s internal data structure. Communication between the two worlds is done via a protocol orthogonal to the message passing mechanism of objects, namely, via events and data flows. This protocol ensures that the information hiding principle at the object-oriented programming level is not violated while constraints can directly access “hidden” data. Furthermore, CODE is built up of distinct elements, or entity types, like constraint, solver, event, data flow. This structure enables that several special purpose constraint solvers can be defined and made to cooperate to solve complex constraint problems. A prototype implementation was built to study the feasibility of CODE. Therefore, the implementation should correspond directly to the conceptual model. To this end, every entity (object, constraint, solver) of the conceptual model is represented by a separate process in the language MANIFOLD. The (concurrent) processes communicate by events and data flows. The implementation serves to validate the conceptual model and to demonstrate that it is a viable way of combining constraints and objects. After the feasibility study, the prototype was discarded. The gained experience was used to build an implementation of the conceptual model for the two research groups. This implementation encompassed a constraint system with multiple solvers and constraint types. The constraint system was built as an object-oriented library that can be linked to the applications in the respective research groups. Special constructs were designed to ensure information hiding among application objects while constraints and solvers have direct access to the object data. CODE manages the complexity of object-oriented constraint solving by defining a communication protocol to allow the two paradigms to cooperate. The prototype implementation demonstrates that CODE can be implemented into a working system. Finally, the implementation of an actual application shows that the model is suitable for the development of object-oriented software.
keywords Computer Graphics; Object Oriented Programming; Constraint Programming
series thesis:PhD
last changed 2003/02/12 22:37

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 76ff
authors Sorby, S.A.
year 1999
title Developing 3-D Spatial Visualization Skills
source Engineering Design Graphics Journal, vol. 63, no. 2 (Spring): 21-32
summary This article brings up the point that 3-D spatial visualization skills are vital to graphics education. Instructors of graphics education, even though they have highly advanced spatial skills, rarely have the proper training on what spatial skills are or how the development of spatial skills takes place. As a result one must try to have a better understanding of spatial abilities. There are many interpretations as to what spatial skills really are and there is in therefore no one universal definition. As a way to better understand spatial abilities, Maier places them into five categories. The categories are spatial perception, spatial visualization, mental rotations, spatial rotations, and spatial orientation. These categories are vast. As a result of their vastness many of the categories overlap. Another step towards better understanding spatial skills involves differentiating how spatial skills are used while completing a task. Tartre makes a classification for how spatial skills are used while performing a task. The spatial skills are either used as spatial visualization that involves mentally moving the object, or as spatial orientation, which involves mentally moving the object. If the task involves spatial visualization then mental rotation can take place, which involves the entire object, or mental transformation can occur, which only involves part of an object. Visual thinking is a way to understand spatial skills. McKim offers the viewpoint that visual thinking occurs by three kinds of imagery. They are what one sees, what one can imagine, and what one can draw. All of these images interact with one another. Spatial skills are developed primarily in three different stages. This can be see be Piaget's theory on development. In the first stage, two dimensional, topological, skills are acquired. In the second stage, an understanding of 3-D objects, projective skills, from different viewpoints is achieved. Finally in the third stage, there is an understanding of area, volume, distance, translation, rotation and reflection, which is combined with projective skills. Spatial skills are evaluated in a variety of ways. There are tests that assess a person's projective skill level. Examples of these would be the Mental Cutting Test and the Differential Aptitude Test: Spatial Relation. Other tests assess mental rotation. Examples of mental rotation tests are the Purdue Spatial Visualization Test and the Mental Rotation Test. Results of these evaluations show mixed results as to whether there are gender differences in spatial skills. In order to enhance spatial skills, one must not only work with 3-D images, but they must also use concrete models and sketching. Overall I thought this article was very informative. It presented the information in a clear and concise manner. I summarized the information that I thought was especially useful for this class. The article really made me think how important it is not only to have spatial skills, but also to have an understanding of them.
series journal paper
last changed 2003/04/23 15:14

_id 44c0
authors Van Leeuwen, Jos P.
year 1999
title Modelling architectural design information by features : an approach to dynamic product modelling for application in architectural design
source Eindhoven University of Technology
summary Architectural design, like many other human activities, benefits more and more from the ongoing development of information and communication technologies. The traditional paper documents for the representation and communication of design are now replaced by digital media. CAD systems have replaced the drawing board and knowledge systems are used to integrate expert knowledge in the design process. Product modelling is one of the most promising approaches in the developments of the last two decades, aiming in the architectural context at the representation and communication of the information related to a building in all its aspects and during its complete life-cycle. However, after studying both the characteristics of the product modelling approach and the characteristics of architectural design, it is concluded in this research project that product modelling does not suffice for support of architectural design. Architectural design is characterised mainly as a problem solving process, involving illdefined problems that require a very dynamic way of dealing with information that concerns both the problem and emerging solutions. Furthermore, architectural design is in many ways an evolutionary process. In short term this is because of the incremental approach to problem solving in design projects; and in long term because of the stylistic development of designers and the continuous developments in the building and construction industry in general. The requirements that are posed by architectural design are concentrated in the keywords extensibility and flexibility of the design informationmodels. Extensibility means that designers can extend conceptual models with definitions that best suit the design concepts they wish to utilise. Flexibility means that information in design models can be structured in a way that accurately represents the design rationale. This includes the modelling of incidental characteristics and relationships of the entities in the model that are not necessarily predefined in a conceptual model. In general, product modelling does not adequately support this dynamic nature of design. Therefore, this research project has studied the concepts developed in the technology of Feature-based modelling, which originates from the area of mechanical engineering. These concepts include the usage of Features as the primitives for defining and reasoning about a product. Features have an autonomous function in the information model, which, as a result, constitutes a flexible network of relationships between Features that are established during the design process. The definition of Features can be specified by designers to formalise new design concepts. This allows the design tools to be adapted to the specific needs of the individual designer, enlarging the library of available resources for design. In addition to these key-concepts in Feature-based modelling as it is developed in the mechanical engineering context, the project has determined the following principles for a Feature-based approach in the architectural context. Features in mechanical engineering are used mainly to describe the lowest level of detail in a product's design, namely the characteristics of its parts. In architecture the design process does not normally follow a strictly hierarchical approach and therefore requires that the building be modelled as a whole. This implies that multiple levels of abstraction are modelled and that Features are used to describe information at the various abstraction levels. Furthermore, architectural design involves concepts that are non-physical as well as physical; Features are to be used for modelling both kinds. The term Feature is defined in this research project to reflect the above key-concepts for this modelling approach. A Feature is an autonomous, coherent collection of information, with semantic meaning to a designer and possibly emerging during design, that is defined to formalise a design concept at any level of abstraction, either physical or non-physical, as part of a building model. Feature models are built up entirely of Features and are structured in the form of a directed graph. The nodes in the graph are the Features, whereas the arcs are the relationships between the Features. Features can be of user-defined types and incidental relationships can be added that are not defined at the typological level. An inventory in this project of what kind of information is involved in the practice of modelling architectural design is based on the analysis of a selection of sources of architectural design information. This inventory is deepened by a case study and results in the proposition of a categorisation of architectural Feature types.
keywords Automated Management Information Systems; Computer Aided Architectural Design; Information Systems; Modelling
series thesis:PhD
email
more http://www.ds.arch.tue.nl/jos/thesis/
last changed 2003/02/12 22:37

_id e719
authors Achten, Henri and Turksma, Arthur
year 1999
title Virtual Reality in Early Design: the Design Studio Experiences
source AVOCAAD Second International Conference [AVOCAAD Conference Proceedings / ISBN 90-76101-02-07] Brussels (Belgium) 8-10 April 1999, pp. 327-335
summary The Design Systems group of the Eindhoven University of Technology started a new kind of design studio teaching. With the use of high-end equipment, students use Virtual Reality from the very start of the design process. Virtual Reality technology up to now was primarily used for giving presentations. We use the same technology in the design process itself by means of reducing the time span in which one gets results in Virtual Reality. The method is based on a very brief cycle of modelling in AutoCAD, assigning materials in 3DStudio Viz, and then making a walkthrough in Virtual Reality in a standard landscape. Due to this cycle, which takes about 15 seconds, the student gets immediate feedback on design decisions which facilitates evaluation of the design in three dimensions much faster than usual. Usually the learning curve of this kind of software is quite steep, but with the use of templates the number of required steps to achieve results is reduced significantly. In this way, the potential of Virtual Reality is not only explored in research projects, but also in education. This paper discusses the general set-up of the design studio and shows how, via short workshops, students acquire knowledge of the cycle in a short time. The paper focuses on the added value of using Virtual Reality technology in this manner: improved spatial reasoning, translation from two-dimensional to three-dimensional representations, and VR feedback on design decisions. It discusses the needs for new design representations in this design environment, and shows how fast feedback in Virtual Reality can improve the spatial design at an early stage of the design process.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id ae61
authors Af Klercker, Jonas
year 1999
title CAAD - Integrated with the First Steps into Architecture
doi https://doi.org/10.52842/conf.ecaade.1999.266
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 266-272
summary How and when should CAAD be introduced in the curriculum of the School of Architecture? This paper begins with some arguments for starting CAAD education at the very beginning. At the School of Architecture in Lund teachers in the first year courses have tried to integrate CAAD with the introduction to architectural concepts and techniques. Traditionally the first year is divided by several subjects running courses separatly without any contact for coordination. From the academic year 96/97 the teachers of Aplied aestetics, Building Science, Architectural design and CAAD have decided to colaborate as much as possible to make the role of our different fields as clear as possible to the students. Therefore integrating CAAD was a natural step in the academic year 98/99. The computer techniques were taught one step in advance so that the students can practise their understanding of the programs in their tasks in the other subjects. The results were surprisingly good! The students have quickly learned to mix the manual and computer techniques to make expressive and interesting visual presentations of their ideas. Some students with antipaty to computers have overcome this handicap. Some interesting observations are discussed.
keywords Curriculum, First Year Studies, Integration, CAAD, Modelling
series eCAADe
email
last changed 2022/06/07 07:54

_id 7da7
authors Benedetti, Cristina and Salvioni, Giulio
year 1999
title The Use of Renewable Resource in Architecture: New Teaching Methodologies
doi https://doi.org/10.52842/conf.ecaade.1999.751
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 751-756
summary The program is organized into four parts. Each is very much connected, both logically and methodologically, so that the unit as a whole consists of a content and method of access that are not divided up. This method is not in a chronological order that simply goes in one direction, rather it allows the user to "refer back", in real time and in different directions. For the simple purpose of explanation, the sections of the program are listed as follows: (-) "Basic information" concerns the basics of bioclimatic and timber architecture. Without this knowledge, the other two sections would be difficult to understand; (-) "Actual buildings throughout the world"; give examples of architectural quality; they concretize the basics of bioclimatic and timber architecture; (-) "Students' Masters Theses", that follow on from the basic information and the learning experience "in the field", and guided by the lecturer, have a critical approach to actual buildings throughout the world. (-) A multimedia data-sheet organized to ensure a clear and straightforward presentation of information about the construction products. It relies on a tab-based navigation interface that gives users access to eight different stacked windows.
keywords Architecture, Multimedia, Timber, Bioclimatic, Classification
series eCAADe
email
last changed 2022/06/07 07:54

_id aef9
id aef9
authors Brown, A., Knight, M. and Berridge, P. (Eds.)
year 1999
title Architectural Computing from Turing to 2000 [Conference Proceedings]
doi https://doi.org/10.52842/conf.ecaade.1999
source eCAADe Conference Proceedings / ISBN 0-9523687-5-7 / Liverpool (UK) 15-17 September 1999, 773 p.
summary The core theme of this book is the idea of looking forward to where research and development in Computer Aided Architectural Design might be heading. The contention is that we can do so most effectively by using the developments that have taken place over the past three or four decades in Computing and Architectural Computing as our reference point; the past informing the future. The genesis of this theme is the fact that a new millennium is about to arrive. If we are ruthlessly objective the year 2000 holds no more significance than any other year; perhaps we should, instead, be preparing for the year 2048 (2k). In fact, whatever the justification, it is now timely to review where we stand in terms of the development of Architectural Computing. This book aims to do that. It is salutary to look back at what writers and researchers have said in the past about where they thought that the developments in computing were taking us. One of the common themes picked up in the sections of this book is the developments that have been spawned by the global linkup that the worldwide web offers us. In the past decade the scale and application of this new medium of communication has grown at a remarkable rate. There are few technological developments that have become so ubiquitous, so quickly. As a consequence there are particular sections in this book on Communication and the Virtual Design Studio which reflect the prominence of this new area, but examples of its application are scattered throughout the book. In 'Computer-Aided Architectural Design' (1977), Bill Mitchell did suggest that computer network accessibility from expensive centralised locations to affordable common, decentralised computing facilities would become more commonplace. But most pundits have been taken by surprise by just how powerful the explosive cocktail of networks, email and hypertext has proven to be. Each of the ingredients is interesting in its own right but together they have presented us with genuinely new ways of working. Perhaps, with foresight we can see what the next new explosive cocktail might be.
series eCAADe
email
more http://www.ecaade.org
last changed 2022/06/07 07:49

_id ga0020
id ga0020
authors Codignola, G.Matteo
year 2000
title [Title missing]
source International Conference on Generative Art
summary This paper is a summary of my last degree in architecture (discussed in December 1999) with Prof. Celestino Soddu and Prof. Enrica Colabella. In this work I had the possibility to reach complexity by a generative approach with the construction of a paradigm that organizes the different codes of project identity. My general objective was to design shape complexity in variable categories : 3d space surfaces, 2d drawings and 2d textures. I was to discover the identity of one of my favourite architects of the 20th century : Antoni Gaudì, by constructing codes relative to shape complexity. I defined my particular objective in the possibility to abduct from Gaudì's imaginary reference the generatives codes that operate in the logical processing I use to create a possible species project. The next step was to verify the exact working of the new generative codes by means of 3d scenaries, that are recognizable as "Antoni Gaudì specie's architecture". Whit project processing on the generative codes and not on a possible resulting shape design, I was able to organize by my general paradigm the attributes of the project's species : different shapes, different attributes (color, scale, proportion), to get to possible and different scenarys, all recognizable by the relative class codes. I chose three examples in Barcellona built during the period 1902 to 1914 : The Parco Guell, Casa Batllò and Casa Milà are the three reference sceneryes that I used to create the generative codes. In the second step I defined different codes that operate in sequence (it is defined in the paradigm) : The generatives codes are only subjective; they are one possible solution of my interpretation of Antoni Gaudì's identity. This codes operate in four differents ways : Geometrical codes for 2d shapes Geometrical codes for interface relations Spatial codes for 3d extrusion of 2d shapes Geometrical codes for 2d and 3d texturing of generated surfaces. By a stratified application of this codes I arrived at one idea for all the generative processes but many different, possible scenaryes, all recognizable in Gaudì's species. So, my final result has made possible sceneryes belonging to related species defined previously. At the end of my research I designed a project by combination : using Antoni Gaudì's generative codes on a new 3d scenary with a shape catalyst : the Frank Lloyd Wright Guggenheim Museum of New York. In this process I created a "hybrid scenary" : a new species of architectural look; a Guggenheim museum planned by Wright with a god pinch of Gaudì.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 5007
authors Elezkurtaj, Tomor and Franck, Georg
year 1999
title Genetic Algorithms in Support of Creative Architectural Design
doi https://doi.org/10.52842/conf.ecaade.1999.645
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 645-651
summary The functions supported by commercial CAAD software are drawing, construction and presentation. Up to now few programs supporting the creative part of architectural problem solving have become available. The grand hopes of symbolic AI to program creative architectural design have been disappointing. In the meantime, methods called referred to as New AI have become available. Such methods includegenetic algorithms (GA). But GA, though successfully applied in other fields of engineering, still waits to be applied broadly in architectural design. A main problem lies in defining function in architecture. It is much harder to define the function of a building than that of a machine. Without specifying the function of the artifact, the fitness function of the design variants participating in the survival game of artificial evolution remains undetermined. It is impossible to fully specify the fitness function of architecture. The approach presented is one of circumventing a full specification through dividing labor between the GA software and its user. The fitness function of architectural ground plans is typically defined in terms only of the proportions of the room to be accommodated and certain topological relations between them. The rest is left to the human designer who interactively intervenes in the evolution game as displayed on the screen.
keywords Genetic Algorithms, Creative Architectural Design
series eCAADe
email
last changed 2022/06/07 07:55

_id 7ad1
authors Giordano, Rubén F. and Tosello, María Elena
year 1999
title Laberinto: Una Biblioteca para la Virtualidad. Reflexiones y Acontecimientos en el Cyberespacio (Labyrinth: A Library for Virtuality. Reflections and Events in Cyberspace)
source III Congreso Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings] Montevideo (Uruguay) September 29th - October 1st 1999, pp. 83-86
summary This project investigates in the limits of the word like only means of structuring of the thought, before the appearance of new paradigms: the multimedias and the ciber-space that have transformed so much the language written as the architectural one causing unpublished situations: 1.) The transformation of a concrete container to other virtual. 2.) The transformation of the design object, of one static material to another that is a process. 3.) The transformation in the traditional ways of thinking (reversible as the formal logic of the mathematics) to new imaginarys epistemologicals. // These non alone events have caused changes in the forms of to know and to communicate the reality but rather the same one suffers a dilation process. We present for their exploration, a road synthesized in some hypotheses that were elaborated with reason of the International Competition of ACADIA 1998: 1.) The new communication systems (cibercomunication) they generate a new territory that should be colonized. This territory this conformed by objects related by infinite bonds (hipertext). 2.) The topographical form is not lineal and sequential, this it is multidirectional and multiradial. The phenomenon of the blow-up and the dilation are the mechanisms with those that the new objects are generated. 3.) These related fields generate interstitial empty spaces where it appears the desire. The interstice like existential space.
series SIGRADI
email
last changed 2016/03/10 09:52

_id ca7b
authors Howes, Jaki
year 1999
title IT or not IT? An Examination of IT Use in an Experimental Multi-disciplinary Teamwork Situation
doi https://doi.org/10.52842/conf.ecaade.1999.370
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 370-373
summary Leeds Metropolitan University is well placed to carry out research into multi-disciplinary team-working, as all the design and construction disciplines are housed in one faculty. Staff have set up an experimental project, TIME IT (Team-working in Multi-disciplinary Environments using IT) which examines ways of working in the design/construction process and how IT is used when there is no commercial pressure. Four groups of four students, one graduate diploma architect, and one final year student from each of Civil Engineering, Construction Management and Quantity Surveying have been working on feasibility studies for projects that are based on completed schemes or have been devised by collaborators in the Construction Industry. Students have been asked to produce a PowerPoint presentation, in up to five working days, of a design scheme, with costs, structural analysis and construction programme. The students are not assessed on the quality of the product, but on their own ability to monitor the process and use of IT. Despite this, aggressive competition evolved between the teams to produce the 'best' design. Five projects were run in the 1998/99 session. A dedicated IT suite has been provided; each group of students had exclusive use of a machine. They were not told how to approach the projects nor when to use the available technology, but were asked to keep the use of paper to a minimum and to keep all their work on the server, so that it could be monitored externally. Not so. They plotted the AO drawings of an existing building that had been provided on the server. They like paper - they can scribble on it, fold it, tear it and throw it at one another.
keywords IT, Multi-disciplinary, Teamwork
series eCAADe
email
last changed 2022/06/07 07:50

_id f2cf
authors Madrazo, Leandro
year 2001
title NETWORKING : a web environment for a collaborative education
source Stellingwerff, Martijn and Verbeke, Johan (Eds.), ACCOLADE - Architecture, Collaboration, Design. Delft University Press (DUP Science) / ISBN 90-407-2216-1 / The Netherlands, pp. 169-182 [Book ordering info: m.c.stellingwerff@bk.tudelft.nl]
summary NETWORKING is a teaching environment developed in close connection with the course "Sistemas de RepresentaciÛn", which has been carried out for the first time at the E.T.S. d'Arquitectura La Salle, Barcelona, in the academic year 1999/00. It is a web-based environment that promotes the exchange of ideas among students and their capacity to work collaboratively. There is a customized NETWORKING environment for each one of the six themes that made up the course: text, shape, object, image, space and light. In this paper we will discuss two of the six environments: OBJECT and SPACE.
series other
last changed 2001/09/14 21:30

_id f02b
authors Mitchell, W.
year 1999
title E-topia: urban life, Jim –but not as we know it
source MIT press
summary The global digital network is not just a delivery system for email, Web pages, and digital television. It is a whole new urban infrastructure--one that will change the forms of our cities as dramatically as railroads, highways, electric power supply, and telephone networks did in the past. In this lucid, invigorating book, William J. Mitchell examines this new infrastructure and its implications for our future daily lives. Picking up where his best-selling City of Bits left off, Mitchell argues that we must extend the definitions of architecture and urban design to encompass virtual places as well as physical ones, and interconnection by means of telecommunication links as well as by pedestrian circulation and mechanized transportation systems. He proposes strategies for the creation of cities that not only will be sustainable but will make economic, social, and cultural sense in an electronically interconnected and global world. The new settlement patterns of the twenty-first century will be characterized by live/work dwellings, 24-hour pedestrian-scale neighborhoods rich in social relationships, and vigorous local community life, complemented by far-flung configurations of electronic meeting places and decentralized production, marketing, and distribution systems. Neither digiphile nor digiphobe, Mitchell advocates the creation of e-topias--cities that work smarter, not harder.
series other
last changed 2003/04/23 15:14

_id cd2e
authors O’Reilly, T.
year 1999
title Hardware, Software and Infoware
source Di Bona, C., Ockman, S. Stone, M.: Open Sources. Voices from the Open Source Revolution, First Edition, Sebastopol, CA: O’Reilly Publishers
summary I was talking with some friends recently, friends who don't own a computer. They were thinking of getting one so they could use Amazon.com to buy books and CDs. Not to use ``the Internet,'' not to use ``the Web,'' but to use Amazon.com. Now, that's the classic definition of a ``killer application'': one that makes someone go out and buy a computer. What's interesting is that the killer application is no longer a desktop productivity application or even a back-office enterprise software system, but an individual web site. And once you start thinking of web sites as applications, you soon come to realize that they represent an entirely new breed, something you might call an ``information application,'' or perhaps even ``infoware.'' Information applications are used to computerize tasks that just couldn't be handled in the old computing model. A few years ago, if you wanted to search a database of a million books, you talked to a librarian, who knew the arcane search syntax of the available computerized search tools and might be able to find what you wanted. If you wanted to buy a book, you went to a bookstore, and looked through its relatively small selection. Now, tens of thousands of people with no specialized training find and buy books online from that million-record database every day. The secret is that computers have come one step closer to the way that people communicate with each other. Web-based applications use plain English to build their interface -- words and pictures, not specialized little controls that acquire meaning only as you learn the software. Traditional software embeds small amounts of information in a lot of software; infoware embeds small amounts of software in a lot of information. The ``actions'' in an infoware product are generally fairly simple: make a choice, buy or sell, enter a small amount of data, and get back a customized result.
series other
last changed 2003/04/23 15:50

_id 5456
authors Quéau, Philippe
year 1999
title The Information Revolution: in search of the common good
source Mendes, C. & Larreta, E.R. (eds) Media and Social Perception, Rio deJaneiro : UNESCO/ ISSC/ EDCAM, 131-154
summary Is globalization really global? * Glocalization. Towards the Common Good. * One Globe, Lots of Tribes and Many Ghettos * Is There a Pilot in the Global Plane? * The market vs. General interest. The Need for Regulation. * Deregulation does not necessarily mean more competition * Sharing the costs of international calls. The case of Internet. * Price setting and regulation for telecommunications * «Proper» competition vs. «Fair» competition * Public and Private. The crucial importance of «Public Domain» * Is Information Society a Meaningful Society? * New cognitive tools for a global citizenship * The role of UNESCO in the Information Society * Globalization and Abstraction * In search of a global meaning Is globalization really global? Glocalization. The Common Good. Buzz phrases like "global village" or "global information society" are misleading. The concept of "global" is not itself global, in other words globalization is not "universal", it does not affect everybody in the same way. Exactly like the very concept of "universality" is not itself "universal". I am referring here to some Asian voices raising concerns about the "occidental" flavour of the "Universal Declaration of Human Rights" and supporting the idea that "Asian values" were not taken properly into account and might even be in contradiction with it (for instance the confucean predominence of the Society over the rights of the Person).
series other
last changed 2003/04/23 15:14

_id 0647
authors Rosenberg, D.
year 1999
title Use Case Driven Object Modeling with UML
source A Practical Approach. Reading, MA: Addison-Wesley
summary Combining some of today's best ideas about customer-driven object-oriented design, Use Case Driven Object Modeling with UML: A Practical Approach shows you how to use Unified Modeling Language (UML) in the real world, keeping with the author's proprietary software design process. The book begins with the genesis of the author's ICONIX Unified Object Modeling Approach, borrowing ideas and strategies from the "three amigos" who invented UML: Grady Booch, James Rumbaugh, and Ivar Jacobson. Throughout this text, the ICONIX method is used to model a stock trading system, with all the relevant UML diagrams, beginning with class definition and use cases. The author's approach to software relies heavily on customer requirements and use case scenarios for which he has a good deal of practical advice. He provides numerous hints for avoiding bogged-down diagrams. After preliminary design, he advocates drilling down into specifics with robustness diagrams, which trace how classes interact with one another. The most detailed design work comes next with sequence diagrams. Subsequent chapters offer tips on project management, implementation, and testing. Throughout this lively and intelligently organized book, the author presents numerous real-world tips (and Top 10 lists) that supply wisdom to his perspective on effective software design.
series other
last changed 2003/04/23 15:14

_id b34d
authors Russell, P., Kohler, N., Forgber, U., Koch, V. and Rügemer, J.
year 1999
title Interactive Representation of Architectural Design: The Virtual Design Studio as an Architectural Graphics Laboratory
doi https://doi.org/10.52842/conf.ecaade.1999.459
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 459-465
summary This paper introduces the Virtual Design Studio (VDS), an internet based design studio environment established by ifib. VDS transfers lessons learned through research projects in the field of Computer Supported Co-operative Work (CSCW) being carried out at ifib into design education. By training for interdisciplinary co-operation within the design process, the students will become better prepared for the flexibility and co-operability required in planning situations. Increasing the communication and co-operation in the planning process can be achieved through the implementation of IT based virtual workspaces. In the design studio setting, this is done through the use of available internet software and technologies. The methodology of the VDS is briefly described including specific assignments intended to focus student investigations into specific areas including the representation of their work using the world wide web. The pedagogical expectations are discussed and anecdotal evidence precedes an general evaluation of the teaching method. The authors postulate that one of the unintended by-products of the studio is the evolution of an effective use of interactivity in the presentation of design concepts, ideas and solutions. A handful of student work is presented to describe the different approaches taken in using the world wide web (WWW) to display project work. A description of the local evolution (VDS specific) of graphical methods and technologies is followed by a comparison with those used in traditional settings. Representation is discussed with focus on the ability of the WWW to replace, augment or corrupt other methods of presentation. The interactive nature of web based presentations induces alterations to the narration of architectural work and can enhance the spatial perception of design space. Space Perception can be enabled through geometrically true VRML representations, the inclusion of auditory sensations, the abstraction of representation through the use of advertising techniques as well as the introduction of non-linear narrative concepts. Examples used by students are shown. A critical assessment of these new representational methods and the place of current new media within the context of architectural representation is discussed.
keywords Virtual Design Studio, Architectural Graphics, Teaching
series eCAADe
email
last changed 2022/06/07 07:56

_id 2e09
authors Wells, Alison
year 1999
title Exploring the Development of the Independent, Electronic, Scholarly Journal
source University of Sheffield, Department of Information Studies
summary This dissertation sets out to examine the extent of independent, electronic scholarly journals, that is, those that are available through the Internet, usually the World Wide Web, free of charge to the reader, and publish academic articles, usually peer reviewed. A list of journals was drawn up using the NewJour archive of journal announcements, supplemented by the World Wide Web Virtual Library and Glasgow University library catalogue. The 387 journals that were found were then categorised in the following categories: Title, URL, originating country, whether it was peer reviewed, organisation or person responsible, broad subject, narrow subject, format, language, whether there was a print version, year of first issue, year of latest issue and number of issues and articles per year. An e-mail survey was also carried out of the 84 electronic journals that had some way of determining their readership, either by having voluntary or compulsory registration, or sending e-mail updates to subscribers. The survey asked for the number of subscribers to each journal, and also the number of hits the Web site received each month It was found that most of the journals were following one of three strategies: (*) a Niche Market strategy - focusing a highly specialised journal to a small audience. (*) a Vanity Publishing strategy - producing journals in a mainstream area, with no external funding. (*) a Commercial strategy - producing journals in a mainstream area, with attempts made to gain external funding through sponsorship and advertising. It was concluded that unless the current market for electronic journals changed in some way, either by journal budgets devolving to academic departments, or some collaborative action between universities, that the future for electronic journals was as part of "one-stop shops" for particular subjects, with a mixture of free and paid for journals (or article servers), together with other related services, controlled by commercial outfits.
series thesis:MSc
more http://panizzi.shef.ac.uk/elecdiss/edl0001/
last changed 2003/02/12 22:37

_id 3cde
authors Alik, B.
year 1999
title A topology construction from line drawings using a uniform plane subdivision technique
source Computer-Aided Design, Vol. 31 (5) (1999) pp. 335-348
summary The paper describes an algorithm for constructing the topology from a set of line segments or polylines. The problem appears for example at land-maps that have been drawnby general-purpose drawing packages or captured from blue-prints by digitalisation. The solution comprises two steps; in the first step inconsistencies in the input data aredetected and removed, and in the second step the topology is constructed. The algorithm for topology construction consists of two phases: determination of a concave hull,and generation of polygons. It is shown that the running-time of the presented algorithm is better than O(n2), where n is the number of input points. Because of a largenumber of geometric elements being expected, the geometric search needed at the first step of the algorithm is speeded up by an acceleration techniquea uniform planesubdivision.
keywords Computational Geometry, Topology Construction, Uniform Space Subdivision
series journal paper
email
last changed 2003/05/15 21:33

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 31HOMELOGIN (you are user _anon_895440 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002