CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References

Hits 1 to 20 of 713

_id avocaad_2001_09
id avocaad_2001_09
authors Yu-Tung Liu, Yung-Ching Yeh, Sheng-Cheng Shih
year 2001
title Digital Architecture in CAD studio and Internet-based competition
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary Architectural design has been changing because of the vast and creative use of computer in different ways. From the viewpoint of designing itself, computer has been used as drawing tools in the latter phase of design (Mitchell 1977; Coyne et al. 1990), presentation and simulation tools in the middle phase (Liu and Bai 2000), and even critical media which triggers creative thinking in the very early phase (Maher et al. 2000; Liu 1999; Won 1999). All the various roles that computer can play have been adopted in a number of professional design corporations and so-called computer-aided design (CAD) studio in schools worldwide (Kvan 1997, 2000; Cheng 1998). The processes and outcomes of design have been continuously developing to capture the movement of the computer age. However, from the viewpoint of social-cultural theories of architecture, the evolvement of design cannot be achieved solely by designers or design processes. Any new idea of design can be accepted socially, culturally and historically only under one condition: The design outcomes could be reviewed and appreciated by critics in the field at the time of its production (Csikszentmihalyi 1986, 1988; Schon and Wiggins 1992; Liu 2000). In other words, aspects of design production (by designers in different design processes) are as critical as those of design appreciation (by critics in different review processes) in the observation of the future trends of architecture.Nevertheless, in the field of architectural design with computer and Internet, that is, so-called computer-aided design computer-mediated design, or internet-based design, most existing studies pay more attentions to producing design in design processes as mentioned above. Relatively few studies focus on how critics act and how they interact with designers in the review processes. Therefore, this study intends to investigate some evolving phenomena of the interaction between design production and appreciation in the environment of computer and Internet.This paper takes a CAD studio and an Internet-based competition as examples. The CAD studio includes 7 master's students and 2 critics, all from the same countries. The Internet-based competition, held in year 2000, includes 206 designers from 43 counties and 26 critics from 11 countries. 3 students and the 2 critics in the CAD studio are the competition participating designers and critics respectively. The methodological steps are as follows: 1. A qualitative analysis: observation and interview of the 3 participants and 2 reviewers who join both the CAD studio and the competition. The 4 analytical criteria are the kinds of presenting media, the kinds of supportive media (such as verbal and gesture/facial data), stages of the review processes, and interaction between the designer and critics. The behavioral data are acquired by recording the design presentation and dialogue within 3 months. 2. A quantitative analysis: statistical analysis of the detailed reviewing data in the CAD studio and the competition. The four 4 analytical factors are the reviewing time, the number of reviewing of the same project, the comparison between different projects, and grades/comments. 3. Both the qualitative and quantitative data are cross analyzed and discussed, based on the theories of design thinking, design production/appreciation, and the appreciative system (Goodman 1978, 1984).The result of this study indicates that the interaction between design production and appreciation during the review processes could differ significantly. The review processes could be either linear or cyclic due to the influences from the kinds of media, the environmental discrepancies between studio and Internet, as well as cognitive thinking/memory capacity. The design production and appreciation seem to be more linear in CAD studio whereas more cyclic in the Internet environment. This distinction coincides with the complementary observations of designing as a linear process (Jones 1970; Simon 1981) or a cyclic movement (Schon and Wiggins 1992). Some phenomena during the two processes are also illustrated in detail in this paper.This study is merely a starting point of the research in design production and appreciation in the computer and network age. The future direction of investigation is to establish a theoretical model for the interaction between design production and appreciation based on current findings. The model is expected to conduct using revised protocol analysis and interviews. The other future research is to explore how design computing creativity emerge from the process of producing and appreciating.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id ga0007
id ga0007
authors Coates, Paul and Miranda, Pablo
year 2000
title Swarm modelling. The use of Swarm Intelligence to generate architectural form
source International Conference on Generative Art
summary .neither the human purposes nor the architect's method are fully known in advance. Consequently, if this interpretation of the architectural problem situation is accepted, any problem-solving technique that relies on explicit problem definition, on distinct goal orientation, on data collection, or even on non-adaptive algorithms will distort the design process and the human purposes involved.' Stanford Anderson, "Problem-Solving and Problem-Worrying". The works concentrates in the use of the computer as a perceptive device, a sort of virtual hand or "sense", capable of prompting an environment. From a set of data that conforms the environment (in this case the geometrical representation of the form of the site) this perceptive device is capable of differentiating and generating distinct patterns in its behavior, patterns that an observer has to interpret as meaningful information. As Nicholas Negroponte explains referring to the project GROPE in his Architecture Machine: 'In contrast to describing criteria and asking the machine to generate physical form, this exercise focuses on generating criteria from physical form.' 'The onlooking human or architecture machine observes what is "interesting" by observing GROPE's behavior rather than by receiving the testimony that this or that is "interesting".' The swarm as a learning device. In this case the work implements a Swarm as a perceptive device. Swarms constitute a paradigm of parallel systems: a multitude of simple individuals aggregate in colonies or groups, giving rise to collaborative behaviors. The individual sensors can't learn, but the swarm as a system can evolve in to more stable states. These states generate distinct patterns, a result of the inner mechanics of the swarm and of the particularities of the environment. The dynamics of the system allows it to learn and adapt to the environment; information is stored in the speed of the sensors (the more collisions, the slower) that acts as a memory. The speed increases in the absence of collisions and so providing the system with the ability to forget, indispensable for differentiation of information and emergence of patterns. The swarm is both a perceptive and a spatial phenomenon. For being able to Interact with an environment an observer requires some sort of embodiment. In the case of the swarm, its algorithms for moving, collision detection, and swarm mechanics conform its perceptive body. The way this body interacts with its environment in the process of learning and differentiation of spatial patterns constitutes also a spatial phenomenon. The enactive space of the Swarm. Enaction, a concept developed by Maturana and Varela for the description of perception in biological terms, is the understanding of perception as the result of the structural coupling of an environment and an observer. Enaction does not address cognition in the currently conventional sense as an internal manipulation of extrinsic 'information' or 'signals', but as the relation between environment and observer and the blurring of their identities. Thus, the space generated by the swarm is an enactive space, a space without explicit description, and an invention of the swarm-environment structural coupling. If we consider a gestalt as 'Some property -such as roundness- common to a set of sense data and appreciated by organisms or artefacts' (Gordon Pask), the swarm is also able to differentiate space 'gestalts' or spaces of some characteristics, such as 'narrowness', or 'fluidness' etc. Implicit surfaces and the wrapping algorithm. One of the many ways of describing this space is through the use of implicit surfaces. An implicit surface may be imagined as an infinitesimally thin band of some measurable quantity such as color, density, temperature, pressure, etc. Thus, an implicit surface consists of those points in three-space that satisfy some particular requirement. This allows as to wrap the regions of space where a difference of quantity has been produced, enclosing the spaces in which some particular events in the history of the Swarm have occurred. The wrapping method allows complex topologies, such as manifoldness in one continuous surface. It is possible to transform the information generated by the swarm in to a landscape that is the result of the particular reading of the site by the swarm. Working in real time. Because of the complex nature of the machine, the only possible way to evaluate the resulting behavior is in real time. For this purpose specific applications had to be developed, using OpenGL for the Windows programming environment. The package consisted on translators from DXF format to a specific format used by these applications and viceversa, the Swarm "engine", a simulated parallel environment, and the Wrapping programs, to generate the implicit surfaces. Different versions of each had been produced, in different stages of development of the work.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id ec4d
authors Croser, J.
year 2001
title GDL Object
source The Architect’s Journal, 14 June 2001, pp. 49-50
summary It is all too common for technology companies to seek a new route to solving the same problem but for the most part the solutions address the effect and not the cause. The good old-fashioned pencil is the perfect example where inventors have sought to design-out the effect of the inherent brittleness of lead. Traditionally different methods of sharpening were suggested and more recently the propelling pencil has reigned king, the lead being supported by the dispensing sleeve thus reducing the likelihood of breakage. Developers convinced by the Single Building Model approach to design development have each embarked on a difficult journey to create an easy to use feature packed application. Unfortunately it seems that the two are not mutually compatible if we are to believe what we see emanating from Technology giants Autodesk in the guise of Architectural Desktop 3. The effect of their development is a feature rich environment but the cost and in this case the cause is a tool which is far from easy to use. However, this is only a small part of a much bigger problem, Interoperability. You see when one designer develops a model with one tool the information is typically locked in that environment. Of course the geometry can be distributed and shared amongst the team for use with their tools but the properties, or as often misquoted, the intelligence is lost along the way. The effect is the technological version of rubble; the cause is the low quality of data-translation available to us. Fortunately there is one company, which is making rapid advancements on the whole issue of collaboration, and data sharing. An old timer (Graphisoft - famous for ArchiCAD) has just donned a smart new suit, set up a new company called GDL Technology and stepped into the ring to do battle, with a difference. The difference is that GDL Technology does not rely on conquering the competition, quite the opposite in fact their success relies upon the continued success of all the major CAD platforms including AutoCAD, MicroStation and ArchiCAD (of course). GDL Technology have created a standard data format for manufacturers called GDL Objects. Product manufacturers such as Velux are now able to develop product libraries using GDL Objects, which can then be placed in a CAD model, or drawing using almost any CAD tool. The product libraries can be stored on the web or on CD giving easy download access to any building industry professional. These objects are created using scripts which makes them tiny for downloading from the web. Each object contains 3 important types of information: · Parametric scale dependant 2d plan symbols · Full 3d geometric data · Manufacturers information such as material, colour and price Whilst manufacturers are racing to GDL Technologies door to sign up, developers and clients are quick to see the benefit too. Porsche are using GDL Objects to manage their brand identity as they build over 300 new showrooms worldwide. Having defined the building style and interior Porsche, in conjunction with the product suppliers, have produced a CD-ROM with all of the selected building components such as cladding, doors, furniture, and finishes. Designing and detailing the various schemes will therefore be as straightforward as using Lego. To ease the process of accessing, sizing and placing the product libraries GDL Technology have developed a product called GDL Object Explorer, a free-standing application which can be placed on the CD with the product libraries. Furthermore, whilst the Object Explorer gives access to the GDL Objects it also enables the user to save the object in one of many file formats including DWG, DGN, DXF, 3DS and even the IAI's IFC. However, if you are an AutoCAD user there is another tool, which has been designed especially for you, it is called the Object Adapter and it works inside of AutoCAD 14 and 2000. The Object Adapter will dynamically convert all GDL Objects to AutoCAD Blocks during placement, which means that they can be controlled with standard AutoCAD commands. Furthermore, each object can be linked to an online document from the manufacturer web site, which is ideal for more extensive product information. Other tools, which have been developed to make the most of the objects, are the Web Plug-in and SalesCAD. The Plug-in enables objects to be dynamically modified and displayed on web pages and Sales CAD is an easy to learn and use design tool for sales teams to explore, develop and cost designs on a Notebook PC whilst sitting in the architects office. All sales quotations are directly extracted from the model and presented in HTML format as a mixture of product images, product descriptions and tables identifying quantities and costs. With full lifecycle information stored in each GDL Object it is no surprise that GDL Technology see their objects as the future for building design. Indeed they are not alone, the IAI have already said that they are going to explore the possibility of associating GDL Objects with their own data sharing format the IFC. So down to the dirty stuff, money and how much it costs? Well, at the risk of sounding like a market trader in Petticoat Lane, "To you guv? Nuffin". That's right as a user of this technology it will cost you nothing! Not a penny, it is gratis, free. The product manufacturer pays for the license to host their libraries on the web or on CD and even then their costs are small costing from as little as 50p for each CD filled with objects. GDL Technology has come up trumps with their GDL Objects. They have developed a new way to solve old problems. If CAD were a pencil then GDL Objects would be ballistic lead, which would never break or loose its point. A much better alternative to the strategy used by many of their competitors who seek to avoid breaking the pencil by persuading the artist not to press down so hard. If you are still reading and you have not already dropped the magazine and run off to find out if your favorite product supplier has already signed up then I suggest you check out the following web sites www.gdlcentral.com and www.gdltechnology.com. If you do not see them there, pick up the phone and ask them why.
series journal paper
email
last changed 2003/04/23 15:14

_id ga0026
id ga0026
authors Ransen, Owen F.
year 2000
title Possible Futures in Computer Art Generation
source International Conference on Generative Art
summary Years of trying to create an "Image Idea Generator" program have convinced me that the perfect solution would be to have an artificial artistic person, a design slave. This paper describes how I came to that conclusion, realistic alternatives, and briefly, how it could possibly happen. 1. The history of Repligator and Gliftic 1.1 Repligator In 1996 I had the idea of creating an “image idea generator”. I wanted something which would create images out of nothing, but guided by the user. The biggest conceptual problem I had was “out of nothing”. What does that mean? So I put aside that problem and forced the user to give the program a starting image. This program eventually turned into Repligator, commercially described as an “easy to use graphical effects program”, but actually, to my mind, an Image Idea Generator. The first release came out in October 1997. In December 1998 I described Repligator V4 [1] and how I thought it could be developed away from simply being an effects program. In July 1999 Repligator V4 won the Shareware Industry Awards Foundation prize for "Best Graphics Program of 1999". Prize winners are never told why they won, but I am sure that it was because of two things: 1) Easy of use 2) Ease of experimentation "Ease of experimentation" means that Repligator does in fact come up with new graphics ideas. Once you have input your original image you can generate new versions of that image simply by pushing a single key. Repligator is currently at version 6, but, apart from adding many new effects and a few new features, is basically the same program as version 4. Following on from the ideas in [1] I started to develop Gliftic, which is closer to my original thoughts of an image idea generator which "starts from nothing". The Gliftic model of images was that they are composed of three components: 1. Layout or form, for example the outline of a mandala is a form. 2. Color scheme, for example colors selected from autumn leaves from an oak tree. 3. Interpretation, for example Van Gogh would paint a mandala with oak tree colors in a different way to Andy Warhol. There is a Van Gogh interpretation and an Andy Warhol interpretation. Further I wanted to be able to genetically breed images, for example crossing two layouts to produce a child layout. And the same with interpretations and color schemes. If I could achieve this then the program would be very powerful. 1.2 Getting to Gliftic Programming has an amazing way of crystalising ideas. If you want to put an idea into practice via a computer program you really have to understand the idea not only globally, but just as importantly, in detail. You have to make hard design decisions, there can be no vagueness, and so implementing what I had decribed above turned out to be a considerable challenge. I soon found out that the hardest thing to do would be the breeding of forms. What are the "genes" of a form? What are the genes of a circle, say, and how do they compare to the genes of the outline of the UK? I wanted the genotype representation (inside the computer program's data) to be directly linked to the phenotype representation (on the computer screen). This seemed to be the best way of making sure that bred-forms would bare some visual relationship to their parents. I also wanted symmetry to be preserved. For example if two symmetrical objects were bred then their children should be symmetrical. I decided to represent shapes as simply closed polygonal shapes, and the "genes" of these shapes were simply the list of points defining the polygon. Thus a circle would have to be represented by a regular polygon of, say, 100 sides. The outline of the UK could easily be represented as a list of points every 10 Kilometers along the coast line. Now for the important question: what do you get when you cross a circle with the outline of the UK? I tried various ways of combining the "genes" (i.e. coordinates) of the shapes, but none of them really ended up producing interesting shapes. And of the methods I used, many of them, applied over several "generations" simply resulted in amorphous blobs, with no distinct family characteristics. Or rather maybe I should say that no single method of breeding shapes gave decent results for all types of images. Figure 1 shows an example of breeding a mandala with 6 regular polygons: Figure 1 Mandala bred with array of regular polygons I did not try out all my ideas, and maybe in the future I will return to the problem, but it was clear to me that it is a non-trivial problem. And if the breeding of shapes is a non-trivial problem, then what about the breeding of interpretations? I abandoned the genetic (breeding) model of generating designs but retained the idea of the three components (form, color scheme, interpretation). 1.3 Gliftic today Gliftic Version 1.0 was released in May 2000. It allows the user to change a form, a color scheme and an interpretation. The user can experiment with combining different components together and can thus home in on an personally pleasing image. Just as in Repligator, pushing the F7 key make the program choose all the options. Unlike Repligator however the user can also easily experiment with the form (only) by pushing F4, the color scheme (only) by pushing F5 and the interpretation (only) by pushing F6. Figures 2, 3 and 4 show some example images created by Gliftic. Figure 2 Mandala interpreted with arabesques   Figure 3 Trellis interpreted with "graphic ivy"   Figure 4 Regular dots interpreted as "sparks" 1.4 Forms in Gliftic V1 Forms are simply collections of graphics primitives (points, lines, ellipses and polygons). The program generates these collections according to the user's instructions. Currently the forms are: Mandala, Regular Polygon, Random Dots, Random Sticks, Random Shapes, Grid Of Polygons, Trellis, Flying Leap, Sticks And Waves, Spoked Wheel, Biological Growth, Chequer Squares, Regular Dots, Single Line, Paisley, Random Circles, Chevrons. 1.5 Color Schemes in Gliftic V1 When combining a form with an interpretation (described later) the program needs to know what colors it can use. The range of colors is called a color scheme. Gliftic has three color scheme types: 1. Random colors: Colors for the various parts of the image are chosen purely at random. 2. Hue Saturation Value (HSV) colors: The user can choose the main hue (e.g. red or yellow), the saturation (purity) of the color scheme and the value (brightness/darkness) . The user also has to choose how much variation is allowed in the color scheme. A wide variation allows the various colors of the final image to depart a long way from the HSV settings. A smaller variation results in the final image using almost a single color. 3. Colors chosen from an image: The user can choose an image (for example a JPG file of a famous painting, or a digital photograph he took while on holiday in Greece) and Gliftic will select colors from that image. Only colors from the selected image will appear in the output image. 1.6 Interpretations in Gliftic V1 Interpretation in Gliftic is best decribed with a few examples. A pure geometric line could be interpreted as: 1) the branch of a tree 2) a long thin arabesque 3) a sequence of disks 4) a chain, 5) a row of diamonds. An pure geometric ellipse could be interpreted as 1) a lake, 2) a planet, 3) an eye. Gliftic V1 has the following interpretations: Standard, Circles, Flying Leap, Graphic Ivy, Diamond Bar, Sparkz, Ess Disk, Ribbons, George Haite, Arabesque, ZigZag. 1.7 Applications of Gliftic Currently Gliftic is mostly used for creating WEB graphics, often backgrounds as it has an option to enable "tiling" of the generated images. There is also a possibility that it will be used in the custom textile business sometime within the next year or two. The real application of Gliftic is that of generating new graphics ideas, and I suspect that, like Repligator, many users will only understand this later. 2. The future of Gliftic, 3 possibilties Completing Gliftic V1 gave me the experience to understand what problems and opportunities there will be in future development of the program. Here I divide my many ideas into three oversimplified possibilities, and the real result may be a mix of two or all three of them. 2.1 Continue the current development "linearly" Gliftic could grow simply by the addition of more forms and interpretations. In fact I am sure that initially it will grow like this. However this limits the possibilities to what is inside the program itself. These limits can be mitigated by allowing the user to add forms (as vector files). The user can already add color schemes (as images). The biggest problem with leaving the program in its current state is that there is no easy way to add interpretations. 2.2 Allow the artist to program Gliftic It would be interesting to add a language to Gliftic which allows the user to program his own form generators and interpreters. In this way Gliftic becomes a "platform" for the development of dynamic graphics styles by the artist. The advantage of not having to deal with the complexities of Windows programming could attract the more adventurous artists and designers. The choice of programming language of course needs to take into account the fact that the "programmer" is probably not be an expert computer scientist. I have seen how LISP (an not exactly easy artificial intelligence language) has become very popular among non programming users of AutoCAD. If, to complete a job which you do manually and repeatedly, you can write a LISP macro of only 5 lines, then you may be tempted to learn enough LISP to write those 5 lines. Imagine also the ability to publish (and/or sell) "style generators". An artist could develop a particular interpretation function, it creates images of a given character which others find appealing. The interpretation (which runs inside Gliftic as a routine) could be offered to interior designers (for example) to unify carpets, wallpaper, furniture coverings for single projects. As Adrian Ward [3] says on his WEB site: "Programming is no less an artform than painting is a technical process." Learning a computer language to create a single image is overkill and impractical. Learning a computer language to create your own artistic style which generates an infinite series of images in that style may well be attractive. 2.3 Add an artificial conciousness to Gliftic This is a wild science fiction idea which comes into my head regularly. Gliftic manages to surprise the users with the images it makes, but, currently, is limited by what gets programmed into it or by pure chance. How about adding a real artifical conciousness to the program? Creating an intelligent artificial designer? According to Igor Aleksander [1] conciousness is required for programs (computers) to really become usefully intelligent. Aleksander thinks that "the line has been drawn under the philosophical discussion of conciousness, and the way is open to sound scientific investigation". Without going into the details, and with great over-simplification, there are roughly two sorts of artificial intelligence: 1) Programmed intelligence, where, to all intents and purposes, the programmer is the "intelligence". The program may perform well (but often, in practice, doesn't) and any learning which is done is simply statistical and pre-programmed. There is no way that this type of program could become concious. 2) Neural network intelligence, where the programs are based roughly on a simple model of the brain, and the network learns how to do specific tasks. It is this sort of program which, according to Aleksander, could, in the future, become concious, and thus usefully intelligent. What could the advantages of an artificial artist be? 1) There would be no need for programming. Presumbably the human artist would dialog with the artificial artist, directing its development. 2) The artificial artist could be used as an apprentice, doing the "drudge" work of art, which needs intelligence, but is, anyway, monotonous for the human artist. 3) The human artist imagines "concepts", the artificial artist makes them concrete. 4) An concious artificial artist may come up with ideas of its own. Is this science fiction? Arthur C. Clarke's 1st Law: "If a famous scientist says that something can be done, then he is in all probability correct. If a famous scientist says that something cannot be done, then he is in all probability wrong". Arthur C Clarke's 2nd Law: "Only by trying to go beyond the current limits can you find out what the real limits are." One of Bertrand Russell's 10 commandments: "Do not fear to be eccentric in opinion, for every opinion now accepted was once eccentric" 3. References 1. "From Ramon Llull to Image Idea Generation". Ransen, Owen. Proceedings of the 1998 Milan First International Conference on Generative Art. 2. "How To Build A Mind" Aleksander, Igor. Wiedenfeld and Nicolson, 1999 3. "How I Drew One of My Pictures: or, The Authorship of Generative Art" by Adrian Ward and Geof Cox. Proceedings of the 1999 Milan 2nd International Conference on Generative Art.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id ga0001
id ga0001
authors Soddu, Celestino
year 2000
title From Forming to Transforming
source International Conference on Generative Art
summary The ancient codes of harmony stem from the human vision of the complexity of nature. They allow us to think the possible, to design it and to perform its realization. The first gesture of every designer is to take, in a new application that is born from a need the opportunity to experiment with a possible harmonic code. And to operate in the evolution of the project so that this code buds and breeds beauty as a mirror of the complexity and wonder of nature. In this design activity, project after project, every architect builds his own code. This is strongly present in diverse ways in every architect. The code of harmony born from the attention of every man to the complexity of nature, manifests itself in interpretation, which is logical and therefore feasible, of the laws of formalization of relationships. Every interpretation is different and belongs to the oneness of every architect. Every interpretative code stems from, and reveals, our approach to the world, our cultural references, our history, our present and the memory of our past. Each idea is born as a representation of the interpretative code that is a cryptic and subjective code, even if it refers as constant to history of man. Generative art is the maximum expression of this human challenge: it traces a code as a reference to the complexity of nature, and it makes it feasible. So man is the craftsman of the possible, according to the laws of the natural harmony. What does a code of the harmony contain? As for all codes it contains some rules that trace certain behaviors. It is not therefore a sequence, a database of events, of forms, but it defines behaviors: the transformations. To choose forms and to put them together is an activity that can also resemble that of a designer, but essentially it is the activity of the client. The designer does not choose forms but operates transformations, because only by doing so can he put a code of harmony into effect. Between transforming and choosing forms one can trace the borderline between architects and clients, between who designs and who chooses the projected objects. This difference must be reconsidered especially today because we are going toward a hybridization in which the client wants to feel himself a designer, even if he only chooses. And the designer, using sophisticated tools, works as chooser between different solutions, in practice as a client. To design, to create through transformations is, however, an activity that takes time. The generative design, building a usable and upgradable code, makes time virtual and therefore allows the architect, even in a speeded-up world as today is, to design and reach levels of complexity that mirror the complexity of nature and its beauty.
series other
email
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id 83cb
authors Telea, Alexandru C.
year 2000
title Visualisation and simulation with object-oriented networks
source Eindhoven University of Technology
summary Among the existing systems, visual programming environments address best these issues. However, producing interactive simulations and visualisations is still a difficult task. This defines the main research objective of this thesis: The development and implementation of concepts and techniques to combine visualisation, simulation, and application construction in an interactive, easy to use, generic environment. The aim is to produce an environment in which the above mentioned activities can be learnt and carried out easily by a researcher. Working with such an environment should decrease the amount of time usually spent in redesigning existing software elements such as graphics interfaces, existing computational modules, and general infrastructure code. Writing new computational components or importing existing ones should be simple and automatic enough to make using the envisaged system an attractive option for a non programmer expert. Besides this, all proven successful elements of an interactive simulation and visualisation environment should be provided, such as visual programming, graphics user interfaces, direct manipulation, and so on. Finally, a large palette of existing scientific computation, data processing, and visualisation components should be integrated in the proposed system. On one hand, this should prove our claims of openness and easy code integration. On the other hand, this should provide the concrete set of tools needed for building a range of scientific applications and visualisations. This thesis is structured as follows. Chapter 2 defines the context of our work. The scientific research environment is presented and partitioned into the three roles of end user, application designer, and component developer. The interactions between these roles and their specific requirements are described and lead to a more precise formulation of our problem statement. Chapter 3 presents the most used architectures for simulation and visualisation systems: the monolithic system, the application library, and the framework. The advantages and disadvantages of these architectural models are then discussed in relation with our problem statement requirements. The main conclusion drawn is that no single existing architectural model suffices, and that what is needed is a combination of the features present in all three models. Chapter 4 introduces the new architectural model we propose, based on the combination of object-orientation in form of the C++ language and dataflow modelling in the new MC++ language. Chapter 5 presents VISSION, an interactive simulation and visualisation environment constructed on the introduced new architectural model, and shows how the usual tasks of application construction, steering, and visualisation are addressed. In chapter 6, the implementation of VISSION’s architectural model is described in terms of its component parts. Chapter 7 presents the applications of VISSION to numerical simulation, while chapter 8 focuses on its visualisation and graphics applications. Finally, chapter 9 concludes the thesis and outlines possible direction for future research.
keywords Computer Visualisation
series thesis:PhD
email
last changed 2003/02/12 22:37

_id avocaad_2001_17
id avocaad_2001_17
authors Ying-Hsiu Huang, Yu-Tung Liu, Cheng-Yuan Lin, Yi-Ting Cheng, Yu-Chen Chiu
year 2001
title The comparison of animation, virtual reality, and scenario scripting in design process
source AVOCAAD - ADDED VALUE OF COMPUTER AIDED ARCHITECTURAL DESIGN, Nys Koenraad, Provoost Tom, Verbeke Johan, Verleye Johan (Eds.), (2001) Hogeschool voor Wetenschap en Kunst - Departement Architectuur Sint-Lucas, Campus Brussel, ISBN 80-76101-05-1
summary Design media is a fundamental tool, which can incubate concrete ideas from ambiguous concepts. Evolved from freehand sketches, physical models to computerized drafting, modeling (Dave, 2000), animations (Woo, et al., 1999), and virtual reality (Chiu, 1999; Klercker, 1999; Emdanat, 1999), different media are used to communicate to designers or users with different conceptual levels¡@during the design process. Extensively employed in design process, physical models help designers in managing forms and spaces more precisely and more freely (Millon, 1994; Liu, 1996).Computerized drafting, models, animations, and VR have gradually replaced conventional media, freehand sketches and physical models. Diversely used in the design process, computerized media allow designers to handle more divergent levels of space than conventional media do. The rapid emergence of computers in design process has ushered in efforts to the visual impact of this media, particularly (Rahman, 1992). He also emphasized the use of computerized media: modeling and animations. Moreover, based on Rahman's study, Bai and Liu (1998) applied a new design media¡Xvirtual reality, to the design process. In doing so, they proposed an evaluation process to examine the visual impact of this new media in the design process. That same investigation pointed towards the facilitative role of the computerized media in enhancing topical comprehension, concept realization, and development of ideas.Computer technology fosters the growth of emerging media. A new computerized media, scenario scripting (Sasada, 2000; Jozen, 2000), markedly enhances computer animations and, in doing so, positively impacts design processes. For the three latest media, i.e., computerized animation, virtual reality, and scenario scripting, the following question arises: What role does visual impact play in different design phases of these media. Moreover, what is the origin of such an impact? Furthermore, what are the similarities and variances of computing techniques, principles of interaction, and practical applications among these computerized media?This study investigates the similarities and variances among computing techniques, interacting principles, and their applications in the above three media. Different computerized media in the design process are also adopted to explore related phenomenon by using these three media in two projects. First, a renewal planning project of the old district of Hsinchu City is inspected, in which animations and scenario scripting are used. Second, the renewal project is compared with a progressive design project for the Hsinchu Digital Museum, as designed by Peter Eisenman. Finally, similarity and variance among these computerized media are discussed.This study also examines the visual impact of these three computerized media in the design process. In computerized animation, although other designers can realize the spatial concept in design, users cannot fully comprehend the concept. On the other hand, other media such as virtual reality and scenario scripting enable users to more directly comprehend what the designer's presentation.Future studies should more closely examine how these three media impact the design process. This study not only provides further insight into the fundamental characteristics of the three computerized media discussed herein, but also enables designers to adopt different media in the design stages. Both designers and users can more fully understand design-related concepts.
series AVOCAAD
email
last changed 2005/09/09 10:48

_id 4cd1
authors Abdelmawla, S., Elnimeiri, M. and Krawczyk, R.
year 2000
title Structural Gizmos
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 115-121
doi https://doi.org/10.52842/conf.acadia.2000.115
summary Architects are visual learners. The Internet has enabled interactive learning tools that can be used to assist in visual thinking of structural concepts, especially at the introductory levels. Here, we propose a visual approach for understanding structures through a series of interactive learning modules, or ’gizmos’. These gizmos, are the tools that the student may use to examine one structural concept at a time. Being interactive, they offer many more possibilities beyond what one static problem can show. The approach aims to enhance students’ visual intuition, and hence understanding of structural concepts and the parameters affecting design. This paper will present selected structural gizmos, how they work, and how they can enhance structural education for architects.
series ACADIA
email
last changed 2022/06/07 07:54

_id 5cba
authors Anders, Peter
year 1999
title Beyond Y2k: A Look at Acadia's Present and Future
source ACADIA Quarterly, vol. 18, no. 1, p. 10
doi https://doi.org/10.52842/conf.acadia.1999.x.o3r
summary The sky may not be falling, but it sure is getting closer. Where will you when the last three zeros of our millennial odometer click into place? Computer scientists tell us that Y2K will bring the world’s computer infrastructure to its knees. Maybe, maybe not. But it is interesting that Y2K is an issue at all. Speculating on the future is simultaneously a magnifying glass for examining our technologies and a looking glass for what we become through them. "The future" is nothing new. Orwell's vision of totalitarian mass media did come true, if only as Madison Avenue rather than Big Brother. Futureboosters of the '50s were convinced that each garage would house a private airplane by the year 2000. But world citizens of the 60's and 70's feared a nuclear catastrophe that would replace the earth with a smoking crater. Others - perhaps more optimistically -predicted that computers were going to drive all our activities by the year 2000. And, in fact, theymay not be far off... The year 2000 is symbolic marker, a point of reflection and assessment. And - as this date is approaching rapidly - this may be a good time to come to grips with who we are and where we want to be.
series ACADIA
email
last changed 2022/06/07 07:49

_id 8e02
authors Brown, A.G.P. and Coenen, F.P.
year 2000
title Spatial reasoning: improving computational efficiency
source Automation in Construction 9 (4) (2000) pp. 361-367
summary When spatial data is analysed the result is often very computer intensive: even by the standards of contemporary technologies, the machine power needed is great and the processing times significant. This is particularly so in 3-D and 4-D scenarios. What we describe here is a technique, which tackles this and associated problems. The technique is founded in the idea of quad-tesseral addressing; a technique, which was originally applied to the analysis of atomic structures. It is based on ideas concerning Hierarchical clustering developed in the 1960s and 1970s to improve data access time [G.M. Morton, A computer oriented geodetic database and a new technique on file sequencing, IBM Canada, 1996.], and on atomic isohedral (same shape) tiling strategies developed in the 1970s and 1980s concerned with group theory [B. Grunbaum, G.C. Shephard, Tilings and Patterns, Freeman, New York, 1987.]. The technique was first suggested as a suitable representation for GIS in the early 1980s when the two strands were brought together and a tesseral arithmetic applied [F.C. Holdroyd, The Geometry of Tiling Hierarchies, Ars Combanitoria 16B (1983) 211–244.; S.B.M. Bell, B.M. Diaz, F.C. Holroyd, M.J.J. Jackson, Spatially referenced methods of processing raster and vector data, Image and Vision Computing 1 (4) (1983) 211–220.; Diaz, S.B.M. Bell, Spatial Data Processing Using Tesseral Methods, Natural Environment Research Council, Swindon, 1986.]. Here, we describe how that technique can equally be applied to the analysis of environmental interaction with built forms. The way in which the technique deals with the problems described is first to linearise the three-dimensional (3-D) space being investigated. Then, the reasoning applied to that space is applied within the same environment as the definition of the problem data. We show, with an illustrative example, how the technique can be applied. The problem then remains of how to visualise the results of the analysis so undertaken. We show how this has been accomplished so that the 3-D space and the results are represented in a way which facilitates rapid interpretation of the analysis, which has been carried out.
series journal paper
more http://www.elsevier.com/locate/autcon
last changed 2003/05/15 21:22

_id 9384
authors Burry, M., Datta, S. and Anson, S.
year 2000
title Introductory Computer Programming as a Means for Extending Spatial and Temporal Understanding
source Eternity, Infinity and Virtuality in Architecture [Proceedings of the 22nd Annual Conference of the Association for Computer-Aided Design in Architecture / 1-880250-09-8] Washington D.C. 19-22 October 2000, pp. 129-135
doi https://doi.org/10.52842/conf.acadia.2000.129
summary Should computer programming be taught within schools of architecture? Incorporating even low-level computer programming within architectural education curricula is a matter of debate but we have found it useful to do so for two reasons: as an introduction or at least a consolidation of the realm of descriptive geometry and in providing an environment for experimenting in morphological time-based change. Mathematics and descriptive geometry formed a significant proportion of architectural education until the end of the 19th century. This proportion has declined in contemporary curricula, possibly at some cost for despite major advances in automated manufacture, Cartesian measurement is still the principal ‘language’ with which to describe building for construction purposes. When computer programming is used as a platform for instruction in logic and spatial representation, the waning interest in mathematics as a basis for spatial description can be readdressed using a left-field approach. Students gain insights into topology, Cartesian space and morphology through programmatic form finding, as opposed to through direct manipulation. In this context, it matters to the architect-programmer how the program operates more than what it does. This paper describes an assignment where students are given a figurative conceptual space comprising the three Cartesian axes with a cube at its centre. Six Phileban solids mark the Cartesian axial limits to the space. Any point in this space represents a hybrid of one, two or three transformations from the central cube towards the various Phileban solids. Students are asked to predict the topological and morphological outcomes of the operations. Through programming, they become aware of morphogenesis and hybridisation. Here we articulate the hypothesis above and report on the outcome from a student group, whose work reveals wider learning opportunities for architecture students in computer programming than conventionally assumed.
series ACADIA
email
last changed 2022/06/07 07:54

_id 1206
authors Cabezas, M., Mariano, C., Mitolo, S. and Oliva, S.
year 1999
title Transformaciones en el Proceso Enseñanza-Aprendizaje de la Geometría Descriptiva con la Apliacación de los Medios Digitales (Transformations in the Teaching/Learning Process of Descriptive Geometry with the Aplplication of Digital Media)
source III Congreso Iberoamericano de Grafico Digital [SIGRADI Conference Proceedings] Montevideo (Uruguay) September 29th - October 1st 1999, pp. 347-348
summary The insert of the digital technologies in the atmosphere Áulico has left generalizing in a significant way. An example constitutes it the high percentage of students that they manifested general knowledge in the software handling in the introductory course of visual communication, as well as the voluntary presentation of practical works developed with digital means. The necessity of an answer to the requirements that arise of the students sinks to the certainty of a pedagogic compatibility among the matter to try and the teaching attended by the personal computer that would increase the Iconidad and the understanding of a topic of certain complexity like it is the geometry of the space. An educational program designed for the teaching of the Sistema Monge whose general characteristics were presented in the II Ibero-American Seminar of Digital Graph and that it will be applied as experience pilot in the course 2000, it will allow us to respond to the following queries: what place it will be given to the educational program in the formation process in connection with the other pedagogic means.
series SIGRADI
email
last changed 2016/03/10 09:47

_id ddssar0008
id ddssar0008
authors Coyne, R., Lee, J. and Ofluoglu, S.
year 2000
title E-commerce and on-line product information
source Timmermans, Harry (Ed.), Fifth Design and Decision Support Systems in Architecture and Urban Planning - Part one: Architecture Proceedings (Nijkerk, the Netherlands)
summary The spread of the Internet has spurred the recent growth in e-commerce. E-commerce impinges on design in many ways, not least in encouraging the development of new tools for the presentation of product information: that is, information about the products that are used in design projects. (In the case of building design this might include information about standard doors, windows and fittings.) In this paper we present our development of three experimental tools for presenting and organising product information. The tools represent stages in the evolution of a prototype we call PLA(id) (Product Library Assistant – Intranet for Designers) involving the use of multimedia authoring, Java, and mobile computing. We discuss our experimental prototypes under several headings derived from what are commonly considered the key features of e-commerce: the marketplace, collaboration (the electronic marketplace as a community), geography (and geographical boundaries), navigation, regulation, integration and media (how in some cases e-commerce provides both the medium of transaction and the medium of delivery). So this account of our experimentation with product information tools provides a way of reviewing issues raised by e-commerce and their importance for design decision support systems.
series DDSS
last changed 2003/08/07 16:36

_id 623c
authors De Oliveira Zandomeneghi, A.L.A., Kejelin Stradiotto, C.R., Moritz Lima, E.K., Silvano Batista, M., Pinto Lincho, P.R., Costa, R., Da Cunha Silveira, S. and Ribas Ulbrich, S.
year 2000
title Construindo o Conhecimento da Hipermídia (Constructing the Hypermedia Knowledge)
source SIGraDi’2000 - Construindo (n)o espacio digital (constructing the digital Space) [4th SIGRADI Conference Proceedings / ISBN 85-88027-02-X] Rio de Janeiro (Brazil) 25-28 september 2000, pp. 286-288
summary The present article reports the experience of a Mastership/Doctorate class in the Production Engineering Postgraduation Program of Santa Catarina Federal University, with the finality of building a prototype of an educational software. The proposal was to facilitate hypermedia teaching, using hypermedia it self and starting from constructivism basements for the making of the project, named Building Hypermedia. The developed program enveloped pedagogically and operationally three phases: a navigation inside a given hypermedia, a feedback and a conscientization of the traveled way, and the possibility of making a singular hypermedia. The study is here presented on its essence, with the report of the work stages, the discrimination of the basic contents and the insertion of the main frames of the software. The objective is to show what results were reached and demonstrate the use and utility of hypermedia, as a computational mechanism with wide and diversified resources, mainly in teaching area.
series SIGRADI
email
last changed 2016/03/10 09:50

_id 389b
authors Do, Ellen Yi-Luen
year 2000
title Sketch that Scene for Me: Creating Virtual Worlds by Freehand Drawing
source Promise and Reality: State of the Art versus State of Practice in Computing for the Design and Planning Process [18th eCAADe Conference Proceedings / ISBN 0-9523687-6-5] Weimar (Germany) 22-24 June 2000, pp. 265-268
doi https://doi.org/10.52842/conf.ecaade.2000.265
summary With the Web people can now view virtual threedimensional worlds and explore virtual space. Increasingly, novice users are interested in creating 3D Web sites. Virtual Reality Modeling Language gained ISO status in 1997, although it is being supplanted by the compatible Java3D API and alternative 3D Web technologies compete. Viewing VRML scenes is relatively straightforward on most hardware platforms and browsers, but currently there are only two ways to create 3D virtual scenes: One is to code the scene directly using VRML. The other is to use existing CAD and modeling software, and save the world in VRML format or convert to VRML from some other format. Both methods are time consuming, cumbersome, and have steep learning curves. Pen-based user interfaces, on the other hand, are for many an easy and intuitive method for graphics input. Not only are people familiar with the look and feel of paper and pencil, novice users also find it less intimidating to draw what they want, where they want it instead of using a complicated tool palette and pull-down menus. Architects and designers use sketches as a primary tool to generate design ideas and to explore alternatives, and numerous computer-based interfaces have played on the concept of "sketch". However, we restrict the notion of sketch to freehand drawing, which we believe helps people to think, to envision, and to recognize properties of the objects with which they are working. SKETCH employs a pen interface to create three-dimensional models, but it uses a simple language of gestures to control a three-dimensional modeler; it does not attempt to interpret freehand drawings. In contrast, our support of 3D world creation using freehand drawing depend on users’ traditional understanding of a floor plan representation. Igarashi et al. used a pen interface to drive browsing in a 3D world, by projecting the user’s marks on the ground plane in the virtual world. Our Sketch-3D project extends this approach, investigating an interface that allows direct interpretation of the drawing marks (what you draw is what you get) and serves as a rapid prototyping tool for creating 3D virtual scenes.
keywords Freehand Sketching, Pen-Based User Interface, Interaction, VRML, Navigation
series eCAADe
email
more http://www.uni-weimar.de/ecaade/
last changed 2022/06/07 07:55

_id 349e
authors Durmisevic, Sanja
year 2002
title Perception Aspects in Underground Spaces using Intelligent Knowledge Modeling
source Delft University of Technology
summary The intensification, combination and transformation are main strategies for future spatial development of the Netherlands, which are stated in the Fifth Bill regarding Spatial Planning. These strategies indicate that in the future, space should be utilized in a more compact and more efficient way requiring, at the same time, re-evaluation of the existing built environment and finding ways to improve it. In this context, the concept of multiple space usage is accentuated, which would focus on intensive 4-dimensional spatial exploration. The underground space is acknowledged as an important part of multiple space usage. In the document 'Spatial Exploration 2000', the underground space is recognized by policy makers as an important new 'frontier' that could provide significant contribution to future spatial requirements.In a relatively short period, the underground space became an important research area. Although among specialists there is appreciation of what underground space could provide for densely populated urban areas, there are still reserved feelings by the public, which mostly relate to the poor quality of these spaces. Many realized underground projects, namely subways, resulted in poor user satisfaction. Today, there is still a significant knowledge gap related to perception of underground space. There is also a lack of detailed documentation on actual applications of the theories, followed by research results and applied techniques. This is the case in different areas of architectural design, but for underground spaces perhaps most evident due to their infancv role in general architectural practice. In order to create better designs, diverse aspects, which are very often of qualitative nature, should be considered in perspective with the final goal to improve quality and image of underground space. In the architectural design process, one has to establish certain relations among design information in advance, to make design backed by sound rationale. The main difficulty at this point is that such relationships may not be determined due to various reasons. One example may be the vagueness of the architectural design data due to linguistic qualities in them. Another, may be vaguely defined design qualities. In this work, the problem was not only the initial fuzziness of the information but also the desired relevancy determination among all pieces of information given. Presently, to determine the existence of such relevancy is more or less a matter of architectural subjective judgement rather than systematic, non-subjective decision-making based on an existing design. This implies that the invocation of certain tools dealing with fuzzy information is essential for enhanced design decisions. Efficient methods and tools to deal with qualitative, soft data are scarce, especially in the architectural domain. Traditionally well established methods, such as statistical analysis, have been used mainly for data analysis focused on similar types to the present research. These methods mainly fall into a category of pattern recognition. Statistical regression methods are the most common approaches towards this goal. One essential drawback of this method is the inability of dealing efficiently with non-linear data. With statistical analysis, the linear relationships are established by regression analysis where dealing with non-linearity is mostly evaded. Concerning the presence of multi-dimensional data sets, it is evident that the assumption of linear relationships among all pieces of information would be a gross approximation, which one has no basis to assume. A starting point in this research was that there maybe both linearity and non-linearity present in the data and therefore the appropriate methods should be used in order to deal with that non-linearity. Therefore, some other commensurate methods were adopted for knowledge modeling. In that respect, soft computing techniques proved to match the quality of the multi-dimensional data-set subject to analysis, which is deemed to be 'soft'. There is yet another reason why soft-computing techniques were applied, which is related to the automation of knowledge modeling. In this respect, traditional models such as Decision Support Systems and Expert Systems have drawbacks. One important drawback is that the development of these systems is a time-consuming process. The programming part, in which various deliberations are required to form a consistent if-then rule knowledge based system, is also a time-consuming activity. For these reasons, the methods and tools from other disciplines, which also deal with soft data, should be integrated into architectural design. With fuzzy logic, the imprecision of data can be dealt with in a similar way to how humans do it. Artificial neural networks are deemed to some extent to model the human brain, and simulate its functions in the form of parallel information processing. They are considered important components of Artificial Intelligence (Al). With neural networks, it is possible to learn from examples, or more precisely to learn from input-output data samples. The combination of the neural and fuzzy approach proved to be a powerful combination for dealing with qualitative data. The problem of automated knowledge modeling is efficiently solved by employment of machine learning techniques. Here, the expertise of prof. dr. Ozer Ciftcioglu in the field of soft computing was crucial for tool development. By combining knowledge from two different disciplines a unique tool could be developed that would enable intelligent modeling of soft data needed for support of the building design process. In this respect, this research is a starting point in that direction. It is multidisciplinary and on the cutting edge between the field of Architecture and the field of Artificial Intelligence. From the architectural viewpoint, the perception of space is considered through relationship between a human being and a built environment. Techniques from the field of Artificial Intelligence are employed to model that relationship. Such an efficient combination of two disciplines makes it possible to extend our knowledge boundaries in the field of architecture and improve design quality. With additional techniques, meta know/edge, or in other words "knowledge about knowledge", can be created. Such techniques involve sensitivity analysis, which determines the amount of dependency of the output of a model (comfort and public safety) on the information fed into the model (input). Another technique is functional relationship modeling between aspects, which is derivation of dependency of a design parameter as a function of user's perceptions. With this technique, it is possible to determine functional relationships between dependent and independent variables. This thesis is a contribution to better understanding of users' perception of underground space, through the prism of public safety and comfort, which was achieved by means of intelligent knowledge modeling. In this respect, this thesis demonstrated an application of ICT (Information and Communication Technology) as a partner in the building design process by employing advanced modeling techniques. The method explained throughout this work is very generic and is possible to apply to not only different areas of architectural design, but also to other domains that involve qualitative data.
keywords Underground Space; Perception; Soft Computing
series thesis:PhD
email
last changed 2003/02/12 22:37

_id ga0024
id ga0024
authors Ferrara, Paolo and Foglia, Gabriele
year 2000
title TEAnO or the computer assisted generation of manufactured aesthetic goods seen as a constrained flux of technological unconsciousness
source International Conference on Generative Art
summary TEAnO (Telematica, Elettronica, Analisi nell'Opificio) was born in Florence, in 1991, at the age of 8, being the direct consequence of years of attempts by a group of computer science professionals to use the digital computers technology to find a sustainable match among creation, generation (or re-creation) and recreation, the three basic keywords underlying the concept of “Littérature potentielle” deployed by Oulipo in France and Oplepo in Italy (see “La Littérature potentielle (Créations Re-créations Récréations) published in France by Gallimard in 1973). During the last decade, TEAnO has been involving in the generation of “artistic goods” in aesthetic domains such as literature, music, theatre and painting. In all those artefacts in the computer plays a twofold role: it is often a tool to generate the good (e.g. an editor to compose palindrome sonnets of to generate antonymic music) and, sometimes it is the medium that makes the fruition of the good possible (e.g. the generator of passages of definition literature). In that sense such artefacts can actually be considered as “manufactured” goods. A great part of such creation and re-creation work has been based upon a rather small number of generation constraints borrowed from Oulipo, deeply stressed by the use of the digital computer massive combinatory power: S+n, edge extraction, phonetic manipulation, re-writing of well known masterpieces, random generation of plots, etc. Regardless this apparently simple underlying generation mechanisms, the systematic use of computer based tools, as weel the analysis of the produced results, has been the way to highlight two findings which can significantly affect the practice of computer based generation of aesthetic goods: ? the deep structure of an aesthetic work persists even through the more “desctructive” manipulations, (such as the antonymic transformation of the melody and lyrics of a music work) and become evident as a sort of profound, earliest and distinctive constraint; ? the intensive flux of computer generated “raw” material seems to confirm and to bring to our attention the existence of what Walter Benjamin indicated as the different way in which the nature talk to a camera and to our eye, and Franco Vaccari called “technological unconsciousness”. Essential references R. Campagnoli, Y. Hersant, “Oulipo La letteratura potenziale (Creazioni Ri-creazioni Ricreazioni)”, 1985 R. Campagnoli “Oupiliana”, 1995 TEAnO, “Quaderno n. 2 Antologia di letteratura potenziale”, 1996 W. Benjiamin, “Das Kunstwerk im Zeitalter seiner technischen Reprodizierbarkeit”, 1936 F. Vaccari, “Fotografia e inconscio tecnologico”, 1994
series other
more http://www.generativeart.com/
last changed 2003/08/07 17:25

_id f586
authors Gabriel, G. and Maher, M.L.
year 2000
title Analysis of design communication with and without computer mediation
source Proceedings of Co-designing 2000, pp. 329-337
summary With recent developments in CAD and communication technologies, the way we visualise and communicate design representations is changing. A matter of great interest to architects, practitioners and researchers alike, is how computer technology might affect the way they think and work. The concern is not about the notion of 'support' alone, but about ensuring that computers do not disrupt the design process and collaborative activity already going on (Bannon and Schmidt, 1991). Designing new collaborative tools will then have to be guided by a better understanding of how collaborative work is accomplished and by understanding what resources the collaborators use and what hindrances they encounter in their work (Finholt et al., 1990). Designing, as a more abstract notion, is different than having a business meeting using video conferencing. In design it is more important to 'see' what is being discussed rather than 'watch' the other person(s) involved in the discussion. In other words the data being conveyed might be of more importance than the method with which it is communicated (See Kvan, 1994). Similarly, we believe that by using text instead of audio as a medium for verbal communication, verbal representations can then be recorded alongside graphical representations for later retrieval and use. In this paper we present the results of a study on collaborative design in three different environments: face-to-face (FTF), computer-mediated using video conferencing (CMCD-a), and computer-mediated using "talk by typing" (CMCD-b). The underlying aim is to establish a clearer notion of the collaborative needs of architects using computer-mediation. In turn this has the potential in assisting developers when designing new collaborative tools and in assisting designers when selecting an environment for a collaborative session.
series other
last changed 2003/04/23 15:50

_id gerardgabriel_phd
id gerardgabriel_phd
authors Gabriel, Gerard Caesar
year 2000
title COMPUTER-MEDIATED COMMUNICATION IN DESIGN
source PhD Thesis, Faculty of Architecture, University of Sydney
summary Up till now, architects collaborating with other colleagues did so mostly face-to-face (FTF). They had to be in the same space (co-located) at the same time. Communication was ‘spontaneous’ and ideas were represented, whether verbal or nonverbal, by talking and using ‘traditional drawing tools’. If they were geographically displaced, the interaction was then space affected as well as the probability of being time affected. In this case communication was usually mediated through the telephone, and graphically represented ideas were sent by Fax or posted documents. Recently, some architectural firms started using modems and Internet connections to exchange information, by transferring CAD drawings as well as design information, through e-mail and file transfer protocol (FTP). Discussing ideas in architecture, as a more abstract notion, is different from discussing other more concrete arguments using video conferencing. It is more important to ‘see’ what is being discussed at hand than ‘watch’ the other person(s) involved in the discussion. In other words the data being conveyed might be of more importance than the mode of communication. Taking into consideration recent developments in computer and communication technologies this thesis investigates different communication channels utilised in architectural collaboration through Computer Mediated Collaborative Design (CMCD) sessions as opposed to FTF sessions. This thesis investigates the possible effects these different channels have on collaborative design in general and collaborative design communication in particular. We argue that successful CMCD does not necessarily mean emulating close proximity environments. Excluding certain communication channels in a CMCD environment might affect the flow and quantity of synchronous collaborative communication, but not necessarily the quality and content of mutually communicated and represented design ideas. Therefore different communication channels might affect the type of communication and not necessarily the content of the communication. We propose that audio and video are not essential communication channels in CMCD environments. We posit that architects will collaborate and communicate design representations effectively although with some differences, since those two channels might cause interruptions and successful collaborative sessions can take place without them. For this purpose we conducted twenty-four one-hour experiments involving final year architecture students all working to the same design brief. The experiments were divided into three categories, FTF, full computer mediated collaborative design sessions (CMCD-a; audio-video conferencing plus whiteboard as a shared drawing space) and limited computer mediated collaborative design sessions (CMCD-b; with Lambda MOO used as a chat medium plus whiteboard as a shared drawing space). The experiments were video and audio taped, transcribed and coded into a custom developed coding scheme. The results of the analysed coded data and observations of the videotapes provided evidence that there were noticeable differences between the three categories. There was more design communication and less communication control in the CMCD-b category compared to the FTF and CMCD-a categories. Verbal communication became shorter and straight to the point in CMCD-b as opposed to spontaneous non-stop chat in the other two categories. Moreover in CMCD-b the subjects were observed to be more reflective as well as choosing and re-examining their words to explain ideas to their partners. At times they were seen scrolling back through the text of the conversation in order to re-analyse or interpret the design ideas at hand. This was impossible in FTF and CMCD-a sessions, since the subjects were more spontaneous and audio representations were lost as soon as they were uttered. Also the video channel in the CMCD-a category was ignored and hardly used except for the first few minutes of the experiments, for a brief exchange of light humour on the appearance of each subject. The results obtained from analysing the experiments helped us conclude that different communication channels produce different collaborative environments. The three categories of communication for architectural collaboration explored in our experiments are indicative of the alternatives available to architects now. What is not clear to architects is why they would choose one category over another. We propose that each category has its own strengths and difficulties for architectural collaboration, and therefore should be selected on the basis of the type of communication considered to be most effective for the stage and tasks of the design project.
series thesis:PhD
type normal paper
email
last changed 2005/09/09 13:02

_id c839
authors Hwang, Jie-Eun
year 2002
title SpaceScope: Developing a Spatial Information Retrieval System - Focused on Apartment Unit Floor Plans -
source Yonsei University, Dept. of Housing & Interior Design
summary This research investigates the spatial information retrieval (IR) in architecture focused on constructing efficient metadata that is crucial for data retrieval. Generally speaking, metadata is ‘structured data about data’ to describe resources especially in a digital format. In this research, metadata is a sort of data object to be useful in searching spatial information. Metadata is also used to describe raw spatial data object as not only attribute data but also content structurally and semantic ally. There are two issues that motivate this research; 1) what is the spatial information – that materializes the intangible space as a data object, and 2) how we can search the information efficiently – the content-based information retrieval. Although knowledge of a building’s spatial content is most important in architecture, there has been no logical method to manage it.

From the viewpoint of content-based retrieval, the researcher analyzes spatial information of a floor plan, with a focus on the apartment unit floor plan common in Korea. Then the metadata items are extracted in a structured manner. To manage the items efficiently, the researcher develops a data model for spatial information according to the concept of the “Structured Floor Plan”. The main object of content to retrieve is a spatial network that consists of nodes of spaces and their linkages. There are two ways to organize the metadata: the traditional index files and the RDF (Resource Description Framework). While the index files are still efficient with computability, the RDF applies greater options to retrieve, such as fuzzy predicates, semantic predicates, and so on. To exploit the metadata, this research shows several possibilities of query operations that present a set of sample queries about L-DK(Living room – Dining room – Kitchen). Implementation of the prototype system is divided into three parts: 1) a modeling module using Vitruvius; 2) an indexing module using MS SQL Server? 2000 in conjunction XML; and 3) a browsing module using the SpaceScope browser.

The future works may consider XML-based databases and a knowledge based query language, such as RQL/XQL, working on such databases. The more specific domain knowledge is involved, the more practical systems would be. Even in architecture, there may be a diverse range of domain knowledge, such as design, building performance, facility management, energy management, post occupied evaluation, historical research, and so on. Also the issue of interface should be investigated in depth, so that it will be adequate to the needs of the architectural field.

keywords Content-based Information Retrieval; Metadata; RDF; XML; Spatial Information; Apartment Floor Plan; Semantics
series thesis:MSc
email
last changed 2003/04/25 07:27

For more results click below:

this is page 0show page 1show page 2show page 3show page 4show page 5... show page 35HOMELOGIN (you are user _anon_21590 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002