CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 658

_id f91f
authors Elezkurtaj, Tomor and Franck, Georg
year 2000
title Geometry and Topology. A User-Interface to Artificial Evolution in Architectural Design
source Promise and Reality: State of the Art versus State of Practice in Computing for the Design and Planning Process [18th eCAADe Conference Proceedings / ISBN 0-9523687-6-5] Weimar (Germany) 22-24 June 2000, pp. 309-312
doi https://doi.org/10.52842/conf.ecaade.2000.309
summary The paper presents a system that supports architectural floor plan design interactively. The method of problem solving implemented is a combination of an evolutionary strategy (ES) and a genetic algorithm (GA). The problem to be solved consists of fitting a number of rooms (n) into an outline by observing functional requirements. The rooms themselves are specified concerning size, function and preferred proportion. The functional requirements entering the fitness functions are expressed in terms of the proportions of the rooms and the neighbourhood relations between them. The system is designed to deal with one of the core problems of computer supported creativity in architecture. For architecture, form not only, but also function is relevant. Without specifying the function that a piece of architecture is supposed to fulfil, it is hard to support its design by computerised methods of problem solving and optimisation. In architecture, however, function relates to comfort, easiness of use, and aesthetics as well. Since it is extraordinary hard, if not impossible, to operationalise aesthetics, computer aided support of creative architectural design is still in its infancy.
keywords New AI, Genetic Algorithms, Artificial Evolution, creative Architectural Design, Interactive Design, Topology
series eCAADe
email
more http://www.uni-weimar.de/ecaade/
last changed 2022/06/07 07:55

_id ga0010
id ga0010
authors Moroni, A., Zuben, F. Von and Manzolli, J.
year 2000
title ArTbitrariness in Music
source International Conference on Generative Art
summary Evolution is now considered not only powerful enough to bring about the biological entities as complex as humans and conciousness, but also useful in simulation to create algorithms and structures of higher levels of complexity than could easily be built by design. In the context of artistic domains, the process of human-machine interaction is analyzed as a good framework to explore creativity and to produce results that could not be obtained without this interaction. When evolutionary computation and other computational intelligence methodologies are involved, every attempt to improve aesthetic judgement we denote as ArTbitrariness, and is interpreted as an interactive iterative optimization process. ArTbitrariness is also suggested as an effective way to produce art through an efficient manipulation of information and a proper use of computational creativity to increase the complexity of the results without neglecting the aesthetic aspects [Moroni et al., 2000]. Our emphasis will be in an approach to interactive music composition. The problem of computer generation of musical material has received extensive attention and a subclass of the field of algorithmic composition includes those applications which use the computer as something in between an instrument, in which a user "plays" through the application's interface, and a compositional aid, which a user experiments with in order to generate stimulating and varying musical material. This approach was adopted in Vox Populi, a hybrid made up of an instrument and a compositional environment. Differently from other systems found in genetic algorithms or evolutionary computation, in which people have to listen to and judge the musical items, Vox Populi uses the computer and the mouse as real-time music controllers, acting as a new interactive computer-based musical instrument. The interface is designed to be flexible for the user to modify the music being generated. It explores evolutionary computation in the context of algorithmic composition and provides a graphical interface that allows to modify the tonal center and the voice range, changing the evolution of the music by using the mouse[Moroni et al., 1999]. A piece of music consists of several sets of musical material manipulated and exposed to the listener, for example pitches, harmonies, rhythms, timbres, etc. They are composed of a finite number of elements and basically, the aim of a composer is to organize those elements in an esthetic way. Modeling a piece as a dynamic system implies a view in which the composer draws trajectories or orbits using the elements of each set [Manzolli, 1991]. Nonlinear iterative mappings are associated with interface controls. In the next page two examples of nonlinear iterative mappings with their resulting musical pieces are shown.The mappings may give rise to attractors, defined as geometric figures that represent the set of stationary states of a non-linear dynamic system, or simply trajectories to which the system is attracted. The relevance of this approach goes beyond music applications per se. Computer music systems that are built on the basis of a solid theory can be coherently embedded into multimedia environments. The richness and specialty of the music domain are likely to initiate new thinking and ideas, which will have an impact on areas such as knowledge representation and planning, and on the design of visual formalisms and human-computer interfaces in general. Above and bellow, Vox Populi interface is depicted, showing two nonlinear iterative mappings with their resulting musical pieces. References [Manzolli, 1991] J. Manzolli. Harmonic Strange Attractors, CEM BULLETIN, Vol. 2, No. 2, 4 -- 7, 1991. [Moroni et al., 1999] Moroni, J. Manzolli, F. Von Zuben, R. Gudwin. Evolutionary Computation applied to Algorithmic Composition, Proceedings of CEC99 - IEEE International Conference on Evolutionary Computation, Washington D. C., p. 807 -- 811,1999. [Moroni et al., 2000] Moroni, A., Von Zuben, F. and Manzolli, J. ArTbitration, Las Vegas, USA: Proceedings of the 2000 Genetic and Evolutionary Computation Conference Workshop Program – GECCO, 143 -- 145, 2000.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 86dc
authors Aouad, G., and Price, A.D.F.
year 1993
title An integrated system to aid the planning of concrete structures: introducing the system
source The Int. Journal of Construction IT1(2), pp.1-14
summary This paper reports on the development at Loughborough University of a CAD-based integrated model to aid the planning of in-situ concrete structures. The system development started after a review of the planning models currently available and after a detailed questionnaire survey undertaken amongst the top UK and US contractors on the current status of planning techniques and information technology. The main aim of this system is to automate the planning process of in-situ concrete structures using data generated by CAD systems. So far, the integration of a CAD system (AutoCAD 10) and a computerized scheduling system (Artemis 2000) has been achieved on a typical IBM-PC. This enables the generation of network plans using AutoCAD which are then automatically transferred to the Artemis system for time and cost analyses.Traditionally, construction planners are faced with many conventional drawings and documents which are used to re-extract information relevant to their planning processes. Such an approach can be very inefficient as it involves data double-handling and is often error prone. In addition, current computerized construction planning applications are little more than the automation of manual formulations of plans. For example, data are fed into the planning system and computations are performed using either CPM (Critical Path Method) or PERT (Programme Evaluation and Review Technique). However, data relating to the planning process such as activity lists, resources requirements and durations are not automatically generated within the system. It would thus seem logical to devise a CAD-based integrated planning model which accepts data in its electronic format and involves some integration of the traditional planning approach. This paper introduces the proposed CAD-based integrated planning model and describes its different components. In addition, it discusses the system functional specifications and summarizes the main benefits and limitations of such a model.
series journal paper
last changed 2003/05/15 21:45

_id ga0007
id ga0007
authors Coates, Paul and Miranda, Pablo
year 2000
title Swarm modelling. The use of Swarm Intelligence to generate architectural form
source International Conference on Generative Art
summary .neither the human purposes nor the architect's method are fully known in advance. Consequently, if this interpretation of the architectural problem situation is accepted, any problem-solving technique that relies on explicit problem definition, on distinct goal orientation, on data collection, or even on non-adaptive algorithms will distort the design process and the human purposes involved.' Stanford Anderson, "Problem-Solving and Problem-Worrying". The works concentrates in the use of the computer as a perceptive device, a sort of virtual hand or "sense", capable of prompting an environment. From a set of data that conforms the environment (in this case the geometrical representation of the form of the site) this perceptive device is capable of differentiating and generating distinct patterns in its behavior, patterns that an observer has to interpret as meaningful information. As Nicholas Negroponte explains referring to the project GROPE in his Architecture Machine: 'In contrast to describing criteria and asking the machine to generate physical form, this exercise focuses on generating criteria from physical form.' 'The onlooking human or architecture machine observes what is "interesting" by observing GROPE's behavior rather than by receiving the testimony that this or that is "interesting".' The swarm as a learning device. In this case the work implements a Swarm as a perceptive device. Swarms constitute a paradigm of parallel systems: a multitude of simple individuals aggregate in colonies or groups, giving rise to collaborative behaviors. The individual sensors can't learn, but the swarm as a system can evolve in to more stable states. These states generate distinct patterns, a result of the inner mechanics of the swarm and of the particularities of the environment. The dynamics of the system allows it to learn and adapt to the environment; information is stored in the speed of the sensors (the more collisions, the slower) that acts as a memory. The speed increases in the absence of collisions and so providing the system with the ability to forget, indispensable for differentiation of information and emergence of patterns. The swarm is both a perceptive and a spatial phenomenon. For being able to Interact with an environment an observer requires some sort of embodiment. In the case of the swarm, its algorithms for moving, collision detection, and swarm mechanics conform its perceptive body. The way this body interacts with its environment in the process of learning and differentiation of spatial patterns constitutes also a spatial phenomenon. The enactive space of the Swarm. Enaction, a concept developed by Maturana and Varela for the description of perception in biological terms, is the understanding of perception as the result of the structural coupling of an environment and an observer. Enaction does not address cognition in the currently conventional sense as an internal manipulation of extrinsic 'information' or 'signals', but as the relation between environment and observer and the blurring of their identities. Thus, the space generated by the swarm is an enactive space, a space without explicit description, and an invention of the swarm-environment structural coupling. If we consider a gestalt as 'Some property -such as roundness- common to a set of sense data and appreciated by organisms or artefacts' (Gordon Pask), the swarm is also able to differentiate space 'gestalts' or spaces of some characteristics, such as 'narrowness', or 'fluidness' etc. Implicit surfaces and the wrapping algorithm. One of the many ways of describing this space is through the use of implicit surfaces. An implicit surface may be imagined as an infinitesimally thin band of some measurable quantity such as color, density, temperature, pressure, etc. Thus, an implicit surface consists of those points in three-space that satisfy some particular requirement. This allows as to wrap the regions of space where a difference of quantity has been produced, enclosing the spaces in which some particular events in the history of the Swarm have occurred. The wrapping method allows complex topologies, such as manifoldness in one continuous surface. It is possible to transform the information generated by the swarm in to a landscape that is the result of the particular reading of the site by the swarm. Working in real time. Because of the complex nature of the machine, the only possible way to evaluate the resulting behavior is in real time. For this purpose specific applications had to be developed, using OpenGL for the Windows programming environment. The package consisted on translators from DXF format to a specific format used by these applications and viceversa, the Swarm "engine", a simulated parallel environment, and the Wrapping programs, to generate the implicit surfaces. Different versions of each had been produced, in different stages of development of the work.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id sigradi2006_e183a
id sigradi2006_e183a
authors Costa Couceiro, Mauro
year 2006
title La Arquitectura como Extensión Fenotípica Humana - Un Acercamiento Basado en Análisis Computacionales [Architecture as human phenotypic extension – An approach based on computational explorations]
source SIGraDi 2006 - [Proceedings of the 10th Iberoamerican Congress of Digital Graphics] Santiago de Chile - Chile 21-23 November 2006, pp. 56-60
summary The study describes some of the aspects tackled within a current Ph.D. research where architectural applications of constructive, structural and organization processes existing in biological systems are considered. The present information processing capacity of computers and the specific software development have allowed creating a bridge between two holistic nature disciplines: architecture and biology. The crossover between those disciplines entails a methodological paradigm change towards a new one based on the dynamical aspects of forms and compositions. Recent studies about artificial-natural intelligence (Hawkins, 2004) and developmental-evolutionary biology (Maturana, 2004) have added fundamental knowledge about the role of the analogy in the creative process and the relationship between forms and functions. The dimensions and restrictions of the Evo-Devo concepts are analyzed, developed and tested by software that combines parametric geometries, L-systems (Lindenmayer, 1990), shape-grammars (Stiny and Gips, 1971) and evolutionary algorithms (Holland, 1975) as a way of testing new architectural solutions within computable environments. It is pondered Lamarck´s (1744-1829) and Weismann (1834-1914) theoretical approaches to evolution where can be found significant opposing views. Lamarck´s theory assumes that an individual effort towards a specific evolutionary goal can cause change to descendents. On the other hand, Weismann defended that the germ cells are not affected by anything the body learns or any ability it acquires during its life, and cannot pass this information on to the next generation; this is called the Weismann barrier. Lamarck’s widely rejected theory has recently found a new place in artificial and natural intelligence researches as a valid explanation to some aspects of the human knowledge evolution phenomena, that is, the deliberate change of paradigms in the intentional research of solutions. As well as the analogy between genetics and architecture (Estévez and Shu, 2000) is useful in order to understand and program emergent complexity phenomena (Hopfield, 1982) for architectural solutions, also the consideration of architecture as a product of a human extended phenotype can help us to understand better its cultural dimension.
keywords evolutionary computation; genetic architectures; artificial/natural intelligence
series SIGRADI
email
last changed 2016/03/10 09:49

_id 349e
authors Durmisevic, Sanja
year 2002
title Perception Aspects in Underground Spaces using Intelligent Knowledge Modeling
source Delft University of Technology
summary The intensification, combination and transformation are main strategies for future spatial development of the Netherlands, which are stated in the Fifth Bill regarding Spatial Planning. These strategies indicate that in the future, space should be utilized in a more compact and more efficient way requiring, at the same time, re-evaluation of the existing built environment and finding ways to improve it. In this context, the concept of multiple space usage is accentuated, which would focus on intensive 4-dimensional spatial exploration. The underground space is acknowledged as an important part of multiple space usage. In the document 'Spatial Exploration 2000', the underground space is recognized by policy makers as an important new 'frontier' that could provide significant contribution to future spatial requirements.In a relatively short period, the underground space became an important research area. Although among specialists there is appreciation of what underground space could provide for densely populated urban areas, there are still reserved feelings by the public, which mostly relate to the poor quality of these spaces. Many realized underground projects, namely subways, resulted in poor user satisfaction. Today, there is still a significant knowledge gap related to perception of underground space. There is also a lack of detailed documentation on actual applications of the theories, followed by research results and applied techniques. This is the case in different areas of architectural design, but for underground spaces perhaps most evident due to their infancv role in general architectural practice. In order to create better designs, diverse aspects, which are very often of qualitative nature, should be considered in perspective with the final goal to improve quality and image of underground space. In the architectural design process, one has to establish certain relations among design information in advance, to make design backed by sound rationale. The main difficulty at this point is that such relationships may not be determined due to various reasons. One example may be the vagueness of the architectural design data due to linguistic qualities in them. Another, may be vaguely defined design qualities. In this work, the problem was not only the initial fuzziness of the information but also the desired relevancy determination among all pieces of information given. Presently, to determine the existence of such relevancy is more or less a matter of architectural subjective judgement rather than systematic, non-subjective decision-making based on an existing design. This implies that the invocation of certain tools dealing with fuzzy information is essential for enhanced design decisions. Efficient methods and tools to deal with qualitative, soft data are scarce, especially in the architectural domain. Traditionally well established methods, such as statistical analysis, have been used mainly for data analysis focused on similar types to the present research. These methods mainly fall into a category of pattern recognition. Statistical regression methods are the most common approaches towards this goal. One essential drawback of this method is the inability of dealing efficiently with non-linear data. With statistical analysis, the linear relationships are established by regression analysis where dealing with non-linearity is mostly evaded. Concerning the presence of multi-dimensional data sets, it is evident that the assumption of linear relationships among all pieces of information would be a gross approximation, which one has no basis to assume. A starting point in this research was that there maybe both linearity and non-linearity present in the data and therefore the appropriate methods should be used in order to deal with that non-linearity. Therefore, some other commensurate methods were adopted for knowledge modeling. In that respect, soft computing techniques proved to match the quality of the multi-dimensional data-set subject to analysis, which is deemed to be 'soft'. There is yet another reason why soft-computing techniques were applied, which is related to the automation of knowledge modeling. In this respect, traditional models such as Decision Support Systems and Expert Systems have drawbacks. One important drawback is that the development of these systems is a time-consuming process. The programming part, in which various deliberations are required to form a consistent if-then rule knowledge based system, is also a time-consuming activity. For these reasons, the methods and tools from other disciplines, which also deal with soft data, should be integrated into architectural design. With fuzzy logic, the imprecision of data can be dealt with in a similar way to how humans do it. Artificial neural networks are deemed to some extent to model the human brain, and simulate its functions in the form of parallel information processing. They are considered important components of Artificial Intelligence (Al). With neural networks, it is possible to learn from examples, or more precisely to learn from input-output data samples. The combination of the neural and fuzzy approach proved to be a powerful combination for dealing with qualitative data. The problem of automated knowledge modeling is efficiently solved by employment of machine learning techniques. Here, the expertise of prof. dr. Ozer Ciftcioglu in the field of soft computing was crucial for tool development. By combining knowledge from two different disciplines a unique tool could be developed that would enable intelligent modeling of soft data needed for support of the building design process. In this respect, this research is a starting point in that direction. It is multidisciplinary and on the cutting edge between the field of Architecture and the field of Artificial Intelligence. From the architectural viewpoint, the perception of space is considered through relationship between a human being and a built environment. Techniques from the field of Artificial Intelligence are employed to model that relationship. Such an efficient combination of two disciplines makes it possible to extend our knowledge boundaries in the field of architecture and improve design quality. With additional techniques, meta know/edge, or in other words "knowledge about knowledge", can be created. Such techniques involve sensitivity analysis, which determines the amount of dependency of the output of a model (comfort and public safety) on the information fed into the model (input). Another technique is functional relationship modeling between aspects, which is derivation of dependency of a design parameter as a function of user's perceptions. With this technique, it is possible to determine functional relationships between dependent and independent variables. This thesis is a contribution to better understanding of users' perception of underground space, through the prism of public safety and comfort, which was achieved by means of intelligent knowledge modeling. In this respect, this thesis demonstrated an application of ICT (Information and Communication Technology) as a partner in the building design process by employing advanced modeling techniques. The method explained throughout this work is very generic and is possible to apply to not only different areas of architectural design, but also to other domains that involve qualitative data.
keywords Underground Space; Perception; Soft Computing
series thesis:PhD
email
last changed 2003/02/12 22:37

_id 2aca
authors Faucher, Didier and Nivet, Marie-Laure
year 2000
title Playing with design intents: integrating physical and urban constraints in CAD
source Automation in Construction 9 (1) (2000) pp. 93-105
summary Our work deals with the exploration of a universe of forms that satisfy some design intents. That is, we substitute a "generate and test" approach for a declarative approach in which an object is created from its properties. In this paper we present an original method that takes into account design intents relative to sunlight, visibility and urban regulation. First of all we study how current CAD tools have considered these properties until now. Our conclusion is that the classical design/simulation/analysis process does not suit design practices, especially in the early stages. We think that an improved CAD system should offer the architect the option of manipulating abstract information such as design intents. We define an intent as a conceptual expression of constraints having an influence on the project. For instance, a visual intent will be stated with no reference to vision geometry: "from this place, I want to see the front of the new building". We show how to represent each of these constraints with a 3D volume associated to some characteristics. If some solutions exist, we are sure that they are included in these volumes. For physical phenomena we compute the volume geometry using the principles of inverse simulation. In the case of urban regulation we apply deduction rules. Design intents are solved by means of geometrical entities that represent openings or obstructions in the project. Computing constraint volumes is a way of guiding the architect in his exploration of solutions. Constraint volumes are new spaces that can restore the link between form and phenomenon in a CAD tool. Our approach offers the designer the possibility of manipulating design intents.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:22

_id ga0012
id ga0012
authors Galanter, Philip
year 2000
title GA2: a Programming Environment for Abstract Generative Fine Art
source International Conference on Generative Art
summary Fine artists looking to use computers to create generative works, especially those artists inclined towards abstraction, often face an uncomfortable choice in the selection of software tools. On the one hand there are a number of commercial and shareware programs available which implement a few techniques in an easy to use GUI environment. Unfortunately such programs often impose a certain look or style and are not terribly versatile or expressive. The other choice seems to be writing code from scratch, in a language such as c or Java. This can be very time consuming as every new work seems to demand a new program, and the artist's ability to write code can seldom keep pace with his ability to imagine new visual ideas. This paper describes a software system created by the author called GA2 which has been implemented in the Matlab software environment. By layering GA2 over Matlab the artist can take advantage of a very mature programming environment which includes extensive mathematical libraries, simple graphics routines, GUI construction tools, built-in help facilities, and command line, batch mode, and GUI modes of interaction. In addition, GA2 is very portable and can run on Macintosh, Windows, and Unix systems with almost no incremental effort for multi-platform support. GA2 is a work in progress and an extension of the completed GA1 environment. It is medium independent, and can be used for all manner of image, animation, and sound production. GA1 includes a complete set of genetic algorithm operations for breeding families of graphical marks, a database function for managing and recalling various genes, a set of statistical operations for creating various distributions of marks on a canvas or animation frame, a unique Markov-chain-likeoperator for generating families of visually similar lines or paths, and a complete L-system implementation. GA2 extends GA1 by adding more generative techniques such as tiling and symmetry operations, Thom's cusp catastrophe, and mechanisms inspired by complexity science notions such as cellular automata, fractals, artificial life, and chaos. All of these techniques are encapulated in genetic representations. This paper is supplemented with examples from the authors art work, and comments on the philosophy behind this method of working, and its relation towards the reinvigoration of abstraction after post-modernism.  
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id c839
authors Hwang, Jie-Eun
year 2002
title SpaceScope: Developing a Spatial Information Retrieval System - Focused on Apartment Unit Floor Plans -
source Yonsei University, Dept. of Housing & Interior Design
summary This research investigates the spatial information retrieval (IR) in architecture focused on constructing efficient metadata that is crucial for data retrieval. Generally speaking, metadata is ‘structured data about data’ to describe resources especially in a digital format. In this research, metadata is a sort of data object to be useful in searching spatial information. Metadata is also used to describe raw spatial data object as not only attribute data but also content structurally and semantic ally. There are two issues that motivate this research; 1) what is the spatial information – that materializes the intangible space as a data object, and 2) how we can search the information efficiently – the content-based information retrieval. Although knowledge of a building’s spatial content is most important in architecture, there has been no logical method to manage it.

From the viewpoint of content-based retrieval, the researcher analyzes spatial information of a floor plan, with a focus on the apartment unit floor plan common in Korea. Then the metadata items are extracted in a structured manner. To manage the items efficiently, the researcher develops a data model for spatial information according to the concept of the “Structured Floor Plan”. The main object of content to retrieve is a spatial network that consists of nodes of spaces and their linkages. There are two ways to organize the metadata: the traditional index files and the RDF (Resource Description Framework). While the index files are still efficient with computability, the RDF applies greater options to retrieve, such as fuzzy predicates, semantic predicates, and so on. To exploit the metadata, this research shows several possibilities of query operations that present a set of sample queries about L-DK(Living room – Dining room – Kitchen). Implementation of the prototype system is divided into three parts: 1) a modeling module using Vitruvius; 2) an indexing module using MS SQL Server? 2000 in conjunction XML; and 3) a browsing module using the SpaceScope browser.

The future works may consider XML-based databases and a knowledge based query language, such as RQL/XQL, working on such databases. The more specific domain knowledge is involved, the more practical systems would be. Even in architecture, there may be a diverse range of domain knowledge, such as design, building performance, facility management, energy management, post occupied evaluation, historical research, and so on. Also the issue of interface should be investigated in depth, so that it will be adequate to the needs of the architectural field.

keywords Content-based Information Retrieval; Metadata; RDF; XML; Spatial Information; Apartment Floor Plan; Semantics
series thesis:MSc
email
last changed 2003/04/25 07:27

_id 625d
authors Liapi, Katherine A.
year 2001
title Geometric Configuration and Graphical Representation of Spherical Tensegrity Networks
source Reinventing the Discourse - How Digital Tools Help Bridge and Transform Research, Education and Practice in Architecture [Proceedings of the Twenty First Annual Conference of the Association for Computer-Aided Design in Architecture / ISBN 1-880250-10-1] Buffalo (New York) 11-14 October 2001, pp. 258-267
doi https://doi.org/10.52842/conf.acadia.2001.258
summary The term “Tensegrity,” that describes mainly a structural concept, is used in building design to address a class of structures with very promising applications in architecture. Tensegrity structures are characterized by almost no separation between structural configuration and formal or architectural expression (Liapi 2000). In the last two decades structural and mechanical aspects in the design of these structures have been successfully addressed, while their intriguing morphology has inspired several artists and architects. Yet, very few real world applications of the tensegrity concept in architecture have been encountered. The geometric and topological complexity of tensegrity structures that is inherent to their structural and mechanical basis may account for significant difficulties in the study of their form and their limited application in building design. In this paper an efficient method for the generation of the geometry of spherical tensegrity networks is presented. The method is based on the integration of CAD tools with Descriptive Geometry procedures and allows designers to resolve and visualize the complex geometry of such structures.
keywords Tensegrity Networks, Visualization, Geometric Configuration
series ACADIA
email
last changed 2022/06/07 07:59

_id 03ad
authors Lottaz, C., Smith, I.F.C., Robert-Nicoud, Y. and Faltings, B.V.
year 2000
title Constraint-based support for negotiation in collaborative design
source Artificial Intelligence in Engineering, Vol: 14, Issue: 3, pp. 261-280.
summary Solution spaces are proposed, instead of single solutions only, to support collaborative tasks during design and construction. Currently, partners involved in construction projects typically assign single values for sub-sets of variables and then proceed, often after tedious negotiations with other partners, to integrate these partial solutions into more complete project descriptions. We suggest the use of constraint solving to express possibly large families of acceptable solutions in order to improve the negotiation process in two ways. On one hand, con ict detection can be performed in an automated manner. Through the constraints collaborators impose, they de ne large unfeasible areas where no solution to the problem at hand can be expected. An emty intersectidon of the solution spaces can thus point at a con ict of design goals of the di erent collaborators at an early stage of the design process. On the other hand, important decision support during negotiation is provided. When a solution space is found, collaborators know during negotiation that they are negotiating about feasible solutions. Negotiation is no longer a means to nd a solution to the problem but it takes place in order to nd a good or the best solution. Since the consistency of the design remains ensured, collaborators are expected to be less restrictive towards innovative ideas during negotiation. Moreover, constraint techniques using explicit representations of solution spaces can provide tools to visualize trade-o s and illustrate the impact of certain decisions on other parameters. Thus decision-making is improved during the negotiation. New algorithms have been developed at EPFL for solving multi-dimensional nonlinear inequality constraints on continuous variables. Together with intuitive user interfaces such constraint-based support leads to better change management and easier implementation of least commitment decision strategies. It is expected that the results of this research can improve both the e ciency of negotiation processes and the quality of the achieved results.
series journal paper
last changed 2003/04/23 15:50

_id 53c6
authors Mardaljevic, John
year 2000
title Daylight Simulation: Validation, Sky Models and Daylight Coefficients
source De Montfort University, Leicester, UK
summary The application of lighting simulation techniques for daylight illuminance modelling in architectural spaces is described in this thesis. The prediction tool used for all the work described here is the Radiance lighting simulation system. An overview of the features and capabilities of the Radiance system is presented. Daylight simulation using the Radiance system is described in some detail. The relation between physical quantities and the lighting simulation parameters is made clear in a series of progressively more complex examples. Effective use of the inter-reflection calculation is described. The illuminance calculation is validated under real sky conditions for a full-size office space. The simulation model used sky luminance patterns that were based directly on measurements. Internal illuminance predictions are compared with measurements for 754 skies that cover a wide range of naturally occurring conditions. The processing of the sky luminance measurements for the lighting simulation is described. The accuracy of the illuminance predictions is shown to be, in the main, comparable with the accuracy of the model input data. There were a number of predictions with low accuracy. Evidence is presented to show that these result from imprecision in the model specification - such as, uncertainty of the circumsolar luminance - rather than the prediction algorithms themselves. Procedures to visualise and reduce illuminance and lighting-related data are presented. The ability of sky models to reproduce measured sky luminance patterns for the purpose of predicting internal illuminance is investigated. Four sky models and two sky models blends are assessed. Predictions of internal illuminance using sky models/blends are compared against those using measured sky luminance patterns. The sky model blends and the Perez All-weather model are shown to perform comparably well. Illuminance predictions using measured skies however were invariably better than those using sky models/blends. Several formulations of the daylight coefficient approach for predicting time varying illuminances are presented. Radiance is used to predict the daylight coefficients from which internal illuminances are derived. The form and magnitude of the daylight coefficients are related to the scene geometry and the discretisation scheme. Internal illuminances are derived for four daylight coefficient formulations based on the measured luminance patterns for the 754 skies. For the best of the formulations, the accuracy of the daylight coefficient derived illuminances is shown to be comparable to that using the standard Radiance calculation method. The use of the daylight coefficient approach to both accurately and efficiently predict hourly internal daylight illuminance levels for an entire year is described. Daylight coefficients are invariant to building orientation for a fixed building configuration. This property of daylight coefficients is exploited to yield hourly internal illuminances for a full year as a function of building orientation. Visual data analysis techniques are used to display and process the massive number of derived illuminances.
series thesis:PhD
email
more http://www.iesd.dmu.ac.uk/~jm/thesis/
last changed 2003/02/12 22:37

_id ga0014
id ga0014
authors McGuire, Kevin
year 2000
title Controlling Chaos: a Simple Deterministic System for Creating Complex Organic Shapes
source International Conference on Generative Art
summary It is difficult and frustrating to create complex organic shapes using the current set of computer graphic programs. One reason is because the geometry of nature is different from that of our tools. Its self-similarity and fine detail are derived from growth processes that are very different from the working process imposed by drawing programs. This mismatch makesit difficult to create natural looking artifacts. Drawing programs provide a palette of shapes that may be manipulated in a variety ways, but the palette is limited and based on a cold Euclidean geometry. Clouds, rivers, and rocks are not lines or circles. Paint programs provide interesting filters and effects, but require great skill and effort. Always, the details must be arduously managed by the artist. This limits the artist's expressive power. Fractals have stunning visual richness, but the artist's techniques are limited to those of choosing colours and searching the fractal space. Genetic algorithms provide a powerful means for exploring a space of variations, but the artist's skill is limited by the very difficult ability to arrive at the correct fitness function. It is hard to get the picture you wanted. Ideally, the artist should have macroscopic control over the creation while leaving the computer to manage the microscopic details. For the result to feel organic, the details should be rich, consistent and varied, cohesive but not repetitious. For the results to be reproducible, the system should be deterministic. For it to be expressive there should be a cause-effect relationship between the actions in the program and change in the resulting picture. Finally, it would be interesting if the way we drew was more closely related to the way things grew. We present a simple drawing program which provides this mixture of macroscopic control with free microscopic detail. Through use of an accretion growth model, the artist controls large scale structure while varied details emerge naturally from senstive dependence in the system. Its algorithms are simple and deterministic, so its results are predictable and reproducible. The overall resulting structure can be anticipated, but it can also surprise. Despite its simplicity, it has been used to generate a surprisingly rich assortment of complex organic looking pictures.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 15e4
authors Sariyildiz, S., Stouffs, R. and Tunçer, B.
year 2000
title Vision on ICT Developments for the Building Sector
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 11-18
doi https://doi.org/10.52842/conf.acadia.2000.011
summary The building sector is entering a new era. Developments in information and communication technology have an impact throughout the entire life cycle of a building, not only from a process and technical point of view but also from a creative design point of view. As a result of developments of advanced modeling software for architectural design, the gap between what the architect can envision and what the building technician or product architect can materialize is enlarging. Internet technology has already started to provide a closer link between the participants in the building process, their activities, knowledge, and information. Concurrent and collaborative engineering will be the future of building practice in respect to efficiency and quality improvement of this sector. The nature of the building process is complex, not only from a communication point of view, but also from the information of the number of participants, the spatial organization and the infrastructure etc. In the near future, soft computing techniques such as artificial neural networks, fuzzy logic, and genetic algorithms will make contributions to the problem solving aspects of the complex design process. This paper provides an overview of these and other future developments of information and communication technology (ICT) within the building sector.
series ACADIA
email
last changed 2022/06/07 07:57

_id 3f51
authors Streich, B., Oxman, R. and Fritz, O.
year 2000
title Computer-Simulated Growth Processes in Urban Planning and Architecture
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 233-237
doi https://doi.org/10.52842/conf.acadia.2000.233
summary Urban structures, developed and grown over a period of time, are created by processes that, due to the number of influential factors, are not longer comprehensible as a whole. Their development is very complex and depends on a big number of reciprocal factors that even architects or planners sometimes cannot recognize the formal, functional and rational processes of thinking behind it. The involved mechanisms however are particularly obvious in historical urban structures that came to exist over a period of centuries. The planned relationships within these conglomerates are governed by nearly indiscernible rules and show similarities in form and shape to living and non-living forms in nature. They are clearly analogous to fractals or systems with chaotic behavior. In the course of the research project “media experimental design”, financed by the German Research Foundation, algorithms are sought that are able to simulate urban analogous structures digitally. To this effect the main rules of growth processes are researched and extracted. Then, by following these rules, virtual structures are developed and shown by using powerful three-dimensional techniques. The developed mechanisms allow urban planning to be process-oriented, interactive and flexible for permanently changing parameters. With an implemented set of rules the computer is able to create a design and to react to changing situations. In several experimental studies structures were successfully generated which have different forms and qualities depending on their set of rules. For example, structures were programmed which are similar to a big city while other look like a village in hilly landscape. Diverse rules and strategies have been used in order to reduce them to shape specific factors. The rules for growth are administered by a specifically developed databank with sophisticated search mechanisms using the Issue-Concept- Form tool as case-based-reasoning method.
keywords Simulation, Urban Growth-Processes, Virtual Reality
series ACADIA
email
last changed 2022/06/07 07:56

_id 6c0a
authors Tserng, H. Ping, Ran, Bin and Russell, Jeffrey S.
year 2000
title Interactive path planning for multi-equipment landfill operations
source Automation in Construction 10 (1) (2000) pp. 155-168
summary A methodology and several algorithms for interactive motion planning are developed for multi-equipment landfill operations in an automated landfill system (ALS). A system for establishing ALS is also proposed in the paper. To develop a multi-truck/multi-compactor ALS, the major problems can be classified into three principal categories: (1) navigation system for multiple devices, (2) job-site geometric model, and (3) instantaneous motion planning and control system for equipment in the work site. To solve the problems from the three categories, this paper will present a methodology to simulate the operational processes of landfill vehicles and equipment in pre-planning a landfill project as well as finding efficient and collision-free motion patterns to control autonomous landfill equipment during the construction phase. Furthermore, by linking this system to a global positioning system (GPS), the efficient traffic routing and collision-free path for each piece of equipment can be calculated by using real-time positional data acquisition in a 3-D geometric model of a landfill site.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:23

_id bfec
authors Tserng, H.P., Ran, B. and Russell, J.S.
year 2001
title Erratum to ""Interactive path planning for multi-equipment landfill operations"" [Autom. Constr. 10 (2000) 155-168]"
source Automation in Construction 10 (4) (2001) pp. 541-541
summary A methodology and several algorithms for interactive motion planning are developed for multi-equipment landfill operations in an automated landfill system (ALS). A system for establishing ALS is also proposed in the paper. To develop a multi-truck/multi-compactor ALS, the major problems can he classified into three principal categories: (1) navigation system for multiple devices, (2) job-site geometric model, and (3) instantaneous motion planning and control system for equipment in the work site. To solve the problems from the three categories, this paper will present a methodology to simulate the operational processes of landfill vehicles and equipment in pre-planning a landfill project as well as finding efficient and collision-free motion patterns to control autonomous landfill equipment during the construction phase. Furthermore, by linking this system to a global positioning system (GPS), the efficient traffic routing and collision-free path for each piece of equipment can he calculated by using real-time positional data acquisition in a 3-D geometric model of a landfill site.
keywords Multi-equipment landfill operations; Automated landfill system; Global positioning system
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/06/02 09:33

_id 4fa1
authors Lee, E., Ida, Y., Woo, S. and Sasada, T.
year 1999
title Environmental Design Using Fractals in Computer Graphics
source Architectural Computing from Turing to 2000 [eCAADe Conference Proceedings / ISBN 0-9523687-5-7] Liverpool (UK) 15-17 September 1999, pp. 533-538
doi https://doi.org/10.52842/conf.ecaade.1999.533
summary Computer graphics have developed efficient techniques for visualisation of the real world. Many of the algorithms have a physical basis, such as computational models for the light and the shadow, models of real objects (buildings, mountains, roads and so on) and the simulation of natural phenomenon. Now computer graphics techniques provide the virtual world with a perception of three dimensions. The concept of the virtual world and its technology have been expanding and intensifying in recent years. Almost everything in the real world has been simulated in virtual world. When it comes to a terrain model, what we need is labour and time. But now it is possible to simulate terrain like the real world using fractals in computer graphics with a very small program and small data set. This study aims to show how to build a real world impression in the virtual world. In this paper the authors suggest a landscape design method and show the results of its application.
keywords Fractals, Polygon-Reduction, Computer Graphics, Virtual World, Collaboration
series eCAADe
last changed 2022/06/07 07:51

_id ga0022
id ga0022
authors Tokui, Nao and Iba, Hitoshi
year 2000
title Music Composition with Interactive Evolutionary Computation
source International Conference on Generative Art
summary Interactive Evolutionary Computation (IEC), i.e., Evolutionary Computation whose fitness function is provided by a user his/herself, has been applied to esthetic areas, such as art, design and music. We cannot necessarily define fitness functions explicitly in these areas. With IEC, however, we can embed the user's implicit preference into the optimization system. This paper describes a new approach to music composition, more precisely the composition of rhythms, by means of IEC. The main feature of our method is to combine Genetic Algorithms (GA) and Genetic Programming (GP). In our system, GA individuals represent short pieces of rhythmic patterns, while GP individuals express how these patterns are arranged in terms of their functions. Both populations are evolved interactively through the user's evaluation. The integration of interactive GA and GP makes it possible to search for musical structures effectively in the vast search space. In this paper, we show how successfully our proposed method can generate attractive musical rhythms. The effectiveness of our system is demonstrated by the evolved rhythm phrases, which are available from our web site as sound files.
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 32HOMELOGIN (you are user _anon_934394 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002