id |
acadia19_16 |
authors |
Hosmer, Tyson; Tigas, Panagiotis |
year |
2019 |
title |
Deep Reinforcement Learning for Autonomous Robotic Tensegrity (ART) |
source |
ACADIA 19:UBIQUITY AND AUTONOMY [Proceedings of the 39th Annual Conference of the Association for Computer Aided Design in Architecture (ACADIA) ISBN 978-0-578-59179-7] (The University of Texas at Austin School of Architecture, Austin, Texas 21-26 October, 2019) pp. 16-29 |
doi |
https://doi.org/10.52842/conf.acadia.2019.016
|
summary |
The research presented in this paper is part of a larger body of emerging research into embedding autonomy in the built environment. We develop a framework for designing and implementing effective autonomous architecture defined by three key properties: situated and embodied agency, facilitated variation, and intelligence.We present a novel application of Deep Reinforcement Learning to learn adaptable behaviours related to autonomous mobility, self-structuring, self-balancing, and spatial reconfiguration. Architectural robotic prototypes are physically developed with principles of embodied agency and facilitated variation. Physical properties and degrees of freedom are applied as constraints in a simulated physics-based environment where our simulation models are trained to achieve multiple objectives in changing environments. This holistic and generalizable approach to aligning deep reinforcement learning with physically reconfigurable robotic assembly systems takes into account both computational design and physical fabrication. Autonomous Robotic Tensegrity (ART) is presented as an extended case study project for developing our methodology. Our computational design system is developed in Unity3D with simulated multi-physics and deep reinforcement learning using Unity’s ML-agents framework. Topological rules of tensegrity are applied to develop assemblies with actuated tensile members. Single units and assemblies are trained for a series of policies using reinforcement learning in single-agent and multi-agent setups. Physical robotic prototypes are built and actuated to test simulated results.
|
series |
ACADIA |
type |
normal paper |
email |
|
full text |
file.pdf (6,844,481 bytes) |
references |
Content-type: text/plain
|
Brand, Stewart. (1995)
How Buildings Learn: What Happens after They’re Built
, Penguin
|
|
|
|
Brooks, Rodney A. (1991)
New Approaches to Robotics
, Science 253 (5025): 1227–32
|
|
|
|
Foerster, Jakob, Richard Y Chen, Maruan Al-Shedivat, Shimon Whiteson, Pieter Abbeel, and Igor Mordatch (2018)
Learning with Opponent-Learning Awareness
, Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems, 122–30
|
|
|
|
Gerhart, John, and Marc Kirschner. (2007)
The Theory of Facilitated Variation
, Proceedings of the National Academy of Sciences 104 (suppl 1): 8582–89
|
|
|
|
Graham, Jason M, Albert B Kao, Dylana A Wilhelm, and Simon Garnier. (2017)
Optimal Construction of Army Ant Living Bridges
, Journal of Theoretical Biology 435: 184–98
|
|
|
|
Ingber, Donald E. (1998)
The Architecture of Life
, Scientific American 278 (1): 48–57
|
|
|
|
Jenett, Ben, and Kenneth Cheung. (2017)
Bill-e: Robotic Platform for Locomotion and Manipulation of Lightweight Space Structures
, 25th AIAA/AHS Adaptive Structures Conference, 1876
|
|
|
|
Kober, Jens, J Andrew Bagnell, and Jan Peters. (2013)
Reinforcement Learning in Robotics: A Survey
, The International Journal of Robotics Research 32 (11): 1238–74
|
|
|
|
Kulkarni, Tejas D, Karthik Narasimhan, Ardavan Saeedi, and Josh Tenenbaum. (2016)
Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation
, Advances Neural Information Processing Systems, 3675–83
|
|
|
|
Lee, Seunghye, and Jaehong Lee. (2016)
A Novel Method for Topology Design of Tensegrity Structures
, Composite Structures 152: 11–19
|
|
|
|
Levin, Stephen M. (2006)
Tensegrity: The New Biomechanics
, Textbook of Musculoskeletal Medicine 9
|
|
|
|
Levine, Sergey, Chelsea Finn, Trevor Darrell, and Pieter Abbeel. (2016)
End-to-End Training of Deep Visuomotor Policies
, The Journal of Machine Learning Research 17 (1): 1334–73
|
|
|
|
Li, Yuxi. (2017)
Deep Reinforcement Learning: An Overview
, ArXiv Preprint ArXiv:1701.07274
|
|
|
|
Lu, Andong. (2017)
Autonomous Assembly as the Fourth Approach to Generic Construction
, Architectural Design 87 (4): 128–33
|
|
|
|
Mao, Hongzi, Mohammad Alizadeh, Ishai Menache, and Srikanth Kandula. (2016)
Resource Management with Deep Reinforcement Learning
, Proceedings of the 15th ACM Workshop on Hot Topics Networks, 50–56
|
|
|
|
Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. (2013)
Playing Atari with Deep Reinforcement Learning
, ArXiv Preprint ArXiv:1312.5602
|
|
|
|
Nagase, K, T Yamashita, and N Kawabata (2016)
On a Connectivity Matrix Formula for Tensegrity Prism Plates
, Mechanics Research Communications 77: 29–43
|
|
|
|
Parter, Merav, Nadav Kashtan, and Uri Alon. (2008)
Facilitated Variation: How Evolution Learns from Past Environments to Generalize to New Environments
, PLoS Computational Biology 4 (11): e1000206
|
|
|
|
Pathak, Deepak, Chris Lu, Trevor Darrell, Phillip Isola, and Alexei A Efros. (2019)
Learning to Control Self-Assembling Morphologies: A Study of Generalization via Modularity
, ArXiv Preprint ArXiv:1902.05546
|
|
|
|
Petersen, Kirstin, and Radhika Nagpal. (2017)
Complex Design by Simple Robots: A Collective Embodied Intelligence Approach to Construction
, Architectural Design 87 (4): 44–49
|
|
|
|
last changed |
2022/06/07 07:50 |
|