[{"_id":"12972","conference":{"name":"EG: Eurographics","end_date":"2023-05-12","location":"Saarbrucken, Germany","start_date":"2023-05-08"},"tmp":{"short":"CC BY-NC-ND (4.0)","name":"Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)","legal_code_url":"https://creativecommons.org/licenses/by-nc-nd/4.0/legalcode","image":"/images/cc_by_nc_nd.png"},"type":"journal_article","article_type":"original","keyword":["embroidery","design","directionality","density","image"],"status":"public","date_updated":"2023-08-01T14:47:05Z","ddc":["004"],"file_date_updated":"2023-05-16T08:28:37Z","department":[{"_id":"BeBi"}],"abstract":[{"text":"Embroidery is a long-standing and high-quality approach to making logos and images on textiles. Nowadays, it can also be performed via automated machines that weave threads with high spatial accuracy. A characteristic feature of the appearance of the threads is a high degree of anisotropy. The anisotropic behavior is caused by depositing thin but long strings of thread. As a result, the stitched patterns convey both color and direction. Artists leverage this anisotropic behavior to enhance pure color images with textures, illusions of motion, or depth cues. However, designing colorful embroidery patterns with prescribed directionality is a challenging task, one usually requiring an expert designer. In this work, we propose an interactive algorithm that generates machine-fabricable embroidery patterns from multi-chromatic images equipped with user-specified directionality fields.We cast the problem of finding a stitching pattern into vector theory. To find a suitable stitching pattern, we extract sources and sinks from the divergence field of the vector field extracted from the input and use them to trace streamlines. We further optimize the streamlines to guarantee a smooth and connected stitching pattern. The generated patterns approximate the color distribution constrained by the directionality field. To allow for further artistic control, the trade-off between color match and directionality match can be interactively explored via an intuitive slider. We showcase our approach by fabricating several embroidery paths.","lang":"eng"}],"oa_version":"Published Version","intvolume":" 42","month":"05","publication_status":"published","publication_identifier":{"issn":["1467-8659"]},"language":[{"iso":"eng"}],"file":[{"creator":"mpiovarc","date_updated":"2023-05-16T08:28:37Z","file_size":24003702,"date_created":"2023-05-16T08:28:37Z","file_name":"Zhenyuan2023.pdf","access_level":"open_access","relation":"main_file","content_type":"application/pdf","file_id":"12974","checksum":"4c188c2be4745467a8790bbf5d6491aa","success":1}],"ec_funded":1,"license":"https://creativecommons.org/licenses/by-nc-nd/4.0/","volume":42,"issue":"2","project":[{"grant_number":"M03319","name":"Perception-Aware Appearance Fabrication","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"},{"name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020"}],"citation":{"ieee":"Z. Liu, M. Piovarci, C. Hafner, R. Charrondiere, and B. Bickel, “Directionality-aware design of embroidery patterns,” Computer Graphics Forum, vol. 42, no. 2. Wiley, pp. 397–409, 2023.","short":"Z. Liu, M. Piovarci, C. Hafner, R. Charrondiere, B. Bickel, Computer Graphics Forum 42 (2023) 397–409.","ama":"Liu Z, Piovarci M, Hafner C, Charrondiere R, Bickel B. Directionality-aware design of embroidery patterns. Computer Graphics Forum. 2023;42(2):397-409. doi:10.1111/cgf.14770 ","apa":"Liu, Z., Piovarci, M., Hafner, C., Charrondiere, R., & Bickel, B. (2023). Directionality-aware design of embroidery patterns. Computer Graphics Forum. Saarbrucken, Germany: Wiley. https://doi.org/10.1111/cgf.14770 ","mla":"Liu, Zhenyuan, et al. “Directionality-Aware Design of Embroidery Patterns.” Computer Graphics Forum, vol. 42, no. 2, Wiley, 2023, pp. 397–409, doi:10.1111/cgf.14770 .","ista":"Liu Z, Piovarci M, Hafner C, Charrondiere R, Bickel B. 2023. Directionality-aware design of embroidery patterns. Computer Graphics Forum. 42(2), 397–409.","chicago":"Liu, Zhenyuan, Michael Piovarci, Christian Hafner, Raphael Charrondiere, and Bernd Bickel. “Directionality-Aware Design of Embroidery Patterns.” Computer Graphics Forum. Wiley, 2023. https://doi.org/10.1111/cgf.14770 ."},"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","external_id":{"isi":["001000062600033"]},"article_processing_charge":"No","author":[{"first_name":"Zhenyuan","id":"70f0d7cf-ae65-11ec-a14f-89dfc5505b19","last_name":"Liu","full_name":"Liu, Zhenyuan","orcid":"0000-0001-9200-5690"},{"id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael","full_name":"Piovarci, Michael","last_name":"Piovarci"},{"first_name":"Christian","id":"400429CC-F248-11E8-B48F-1D18A9856A87","last_name":"Hafner","full_name":"Hafner, Christian"},{"last_name":"Charrondiere","full_name":"Charrondiere, Raphael","first_name":"Raphael","id":"a3a24133-2cc7-11ec-be88-8ddaf6f464b1"},{"first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel"}],"title":"Directionality-aware design of embroidery patterns","acknowledgement":"This work was supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (grant agreement No 715767 – MATERIALIZABLE), and FWF Lise Meitner (Grant M 3319). We thank the anonymous reviewers for their insightful feedback; Solal Pirelli, Shardul Chiplunkar, and Paola Mejia for proofreading; everyone in the visual computing group at ISTA for inspiring lunch and coffee breaks; Thibault Tricard for help producing the results of Phasor Noise.","oa":1,"quality_controlled":"1","publisher":"Wiley","year":"2023","has_accepted_license":"1","isi":1,"publication":"Computer Graphics Forum","day":"08","page":"397-409","date_created":"2023-05-16T08:47:25Z","doi":"10.1111/cgf.14770 ","date_published":"2023-05-08T00:00:00Z"},{"oa":1,"quality_controlled":"1","publisher":"Association for Computing Machinery","acknowledgement":"The authors would like to thank Yuki Koyama and Takeo Igarashi for early discussions, and Yuta Yaguchi for support in 3D printing. This research is partially supported by the Israel Science Foundation grant number 1390/19.\r\n","date_created":"2023-08-27T22:01:17Z","date_published":"2023-07-23T00:00:00Z","doi":"10.1145/3588432.3591542","publication":"SIGGRAPH 2023 Conference Proceedings","day":"23","year":"2023","article_number":"20","title":"Stealth shaper: Reflectivity optimization as surface stylization","article_processing_charge":"No","external_id":{"arxiv":["2305.05944"]},"author":[{"last_name":"Tojo","full_name":"Tojo, Kenji","first_name":"Kenji"},{"full_name":"Shamir, Ariel","last_name":"Shamir","first_name":"Ariel"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd","last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385"},{"full_name":"Umetani, Nobuyuki","last_name":"Umetani","first_name":"Nobuyuki"}],"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"chicago":"Tojo, Kenji, Ariel Shamir, Bernd Bickel, and Nobuyuki Umetani. “Stealth Shaper: Reflectivity Optimization as Surface Stylization.” In SIGGRAPH 2023 Conference Proceedings. Association for Computing Machinery, 2023. https://doi.org/10.1145/3588432.3591542.","ista":"Tojo K, Shamir A, Bickel B, Umetani N. 2023. Stealth shaper: Reflectivity optimization as surface stylization. SIGGRAPH 2023 Conference Proceedings. SIGGRAPH: Computer Graphics and Interactive Techniques Conference, 20.","mla":"Tojo, Kenji, et al. “Stealth Shaper: Reflectivity Optimization as Surface Stylization.” SIGGRAPH 2023 Conference Proceedings, 20, Association for Computing Machinery, 2023, doi:10.1145/3588432.3591542.","apa":"Tojo, K., Shamir, A., Bickel, B., & Umetani, N. (2023). Stealth shaper: Reflectivity optimization as surface stylization. In SIGGRAPH 2023 Conference Proceedings. Los Angeles, CA, United States: Association for Computing Machinery. https://doi.org/10.1145/3588432.3591542","ama":"Tojo K, Shamir A, Bickel B, Umetani N. Stealth shaper: Reflectivity optimization as surface stylization. In: SIGGRAPH 2023 Conference Proceedings. Association for Computing Machinery; 2023. doi:10.1145/3588432.3591542","short":"K. Tojo, A. Shamir, B. Bickel, N. Umetani, in:, SIGGRAPH 2023 Conference Proceedings, Association for Computing Machinery, 2023.","ieee":"K. Tojo, A. Shamir, B. Bickel, and N. Umetani, “Stealth shaper: Reflectivity optimization as surface stylization,” in SIGGRAPH 2023 Conference Proceedings, Los Angeles, CA, United States, 2023."},"month":"07","main_file_link":[{"url":"https://doi.org/10.48550/arXiv.2305.05944","open_access":"1"}],"scopus_import":"1","oa_version":"Preprint","abstract":[{"lang":"eng","text":"We present a technique to optimize the reflectivity of a surface while preserving its overall shape. The naïve optimization of the mesh vertices using the gradients of reflectivity simulations results in undesirable distortion. In contrast, our robust formulation optimizes the surface normal as an independent variable that bridges the reflectivity term with differential rendering, and the regularization term with as-rigid-as-possible elastic energy. We further adaptively subdivide the input mesh to improve the convergence. Consequently, our method can minimize the retroreflectivity of a wide range of input shapes, resulting in sharply creased shapes ubiquitous among stealth aircraft and Sci-Fi vehicles. Furthermore, by changing the reward for the direction of the outgoing light directions, our method can be applied to other reflectivity design tasks, such as the optimization of architectural walls to concentrate light in a specific region. We have tested the proposed method using light-transport simulations and real-world 3D-printed objects."}],"language":[{"iso":"eng"}],"publication_status":"published","publication_identifier":{"isbn":["9798400701597"]},"status":"public","conference":{"start_date":"2023-08-06","end_date":"2023-08-10","location":"Los Angeles, CA, United States","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference"},"type":"conference","_id":"14241","department":[{"_id":"BeBi"}],"date_updated":"2023-09-05T07:22:03Z"},{"abstract":[{"lang":"eng","text":"Portrait viewpoint and illumination editing is an important problem with several applications in VR/AR, movies, and photography. Comprehensive knowledge of geometry and illumination is critical for obtaining photorealistic results. Current methods are unable to explicitly model in 3D while handling both viewpoint and illumination editing from a single image. In this paper, we propose VoRF, a novel approach that can take even a single portrait image as input and relight human heads under novel illuminations that can be viewed from arbitrary viewpoints. VoRF represents a human head as a continuous volumetric field and learns a prior model of human heads using a coordinate-based MLP with individual latent spaces for identity and illumination. The prior model is learned in an auto-decoder manner over a diverse class of head shapes and appearances, allowing VoRF to generalize to novel test identities from a single input image. Additionally, VoRF has a reflectance MLP that uses the intermediate features of the prior model for rendering One-Light-at-A-Time (OLAT) images under novel views. We synthesize novel illuminations by combining these OLAT images with target environment maps. Qualitative and quantitative evaluations demonstrate the effectiveness of VoRF for relighting and novel view synthesis, even when applied to unseen subjects under uncontrolled illumination. This work is an extension of Rao et al. (VoRF: Volumetric Relightable Faces 2022). We provide extensive evaluation and ablative studies of our model and also provide an application, where any face can be relighted using textual input."}],"acknowledgement":"Open Access funding enabled and organized by Projekt DEAL.","oa_version":"Published Version","oa":1,"main_file_link":[{"url":"https://doi.org/10.1007/s11263-023-01899-3","open_access":"1"}],"publisher":"Springer Nature","quality_controlled":"1","scopus_import":"1","month":"10","publication_status":"epub_ahead","year":"2023","publication_identifier":{"eissn":["1573-1405"],"issn":["0920-5691"]},"publication":"International Journal of Computer Vision","language":[{"iso":"eng"}],"day":"31","date_created":"2023-11-05T23:00:54Z","date_published":"2023-10-31T00:00:00Z","doi":"10.1007/s11263-023-01899-3","_id":"14488","type":"journal_article","article_type":"original","status":"public","citation":{"ama":"Rao P, Mallikarjun BR, Fox G, et al. A deeper analysis of volumetric relightiable faces. International Journal of Computer Vision. 2023. doi:10.1007/s11263-023-01899-3","apa":"Rao, P., Mallikarjun, B. R., Fox, G., Weyrich, T., Bickel, B., Pfister, H., … Elgharib, M. (2023). A deeper analysis of volumetric relightiable faces. International Journal of Computer Vision. Springer Nature. https://doi.org/10.1007/s11263-023-01899-3","short":"P. Rao, B.R. Mallikarjun, G. Fox, T. Weyrich, B. Bickel, H. Pfister, W. Matusik, F. Zhan, A. Tewari, C. Theobalt, M. Elgharib, International Journal of Computer Vision (2023).","ieee":"P. Rao et al., “A deeper analysis of volumetric relightiable faces,” International Journal of Computer Vision. Springer Nature, 2023.","mla":"Rao, Pramod, et al. “A Deeper Analysis of Volumetric Relightiable Faces.” International Journal of Computer Vision, Springer Nature, 2023, doi:10.1007/s11263-023-01899-3.","ista":"Rao P, Mallikarjun BR, Fox G, Weyrich T, Bickel B, Pfister H, Matusik W, Zhan F, Tewari A, Theobalt C, Elgharib M. 2023. A deeper analysis of volumetric relightiable faces. International Journal of Computer Vision.","chicago":"Rao, Pramod, B. R. Mallikarjun, Gereon Fox, Tim Weyrich, Bernd Bickel, Hanspeter Pfister, Wojciech Matusik, et al. “A Deeper Analysis of Volumetric Relightiable Faces.” International Journal of Computer Vision. Springer Nature, 2023. https://doi.org/10.1007/s11263-023-01899-3."},"date_updated":"2023-11-06T08:52:30Z","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"Yes (via OA deal)","author":[{"first_name":"Pramod","full_name":"Rao, Pramod","last_name":"Rao"},{"first_name":"B. R.","last_name":"Mallikarjun","full_name":"Mallikarjun, B. R."},{"first_name":"Gereon","last_name":"Fox","full_name":"Fox, Gereon"},{"first_name":"Tim","full_name":"Weyrich, Tim","last_name":"Weyrich"},{"first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385"},{"last_name":"Pfister","full_name":"Pfister, Hanspeter","first_name":"Hanspeter"},{"first_name":"Wojciech","full_name":"Matusik, Wojciech","last_name":"Matusik"},{"last_name":"Zhan","full_name":"Zhan, Fangneng","first_name":"Fangneng"},{"full_name":"Tewari, Ayush","last_name":"Tewari","first_name":"Ayush"},{"first_name":"Christian","last_name":"Theobalt","full_name":"Theobalt, Christian"},{"full_name":"Elgharib, Mohamed","last_name":"Elgharib","first_name":"Mohamed"}],"title":"A deeper analysis of volumetric relightiable faces","department":[{"_id":"BeBi"}]},{"date_updated":"2023-12-04T08:09:05Z","ddc":["531","006"],"department":[{"_id":"GradSch"},{"_id":"ChWo"},{"_id":"BeBi"}],"file_date_updated":"2023-12-04T08:04:14Z","_id":"14628","article_type":"original","type":"journal_article","status":"public","keyword":["Computer Graphics and Computer-Aided Design"],"publication_identifier":{"issn":["0730-0301","1557-7368"]},"publication_status":"published","file":[{"date_updated":"2023-11-29T15:16:01Z","file_size":95467870,"creator":"yichen","date_created":"2023-11-29T15:16:01Z","file_name":"tog-22-0089-File004.zip","content_type":"application/zip","access_level":"open_access","relation":"main_file","checksum":"0192f597d7a2ceaf89baddfd6190d4c8","file_id":"14630","success":1},{"checksum":"7fb024963be81933494f38de191e4710","file_id":"14631","success":1,"access_level":"open_access","relation":"main_file","content_type":"application/zip","date_created":"2023-11-29T15:16:01Z","file_name":"tog-22-0089-File005.zip","creator":"yichen","date_updated":"2023-11-29T15:16:01Z","file_size":103731880},{"content_type":"application/pdf","access_level":"open_access","relation":"main_file","file_id":"14638","checksum":"b7d6829ce396e21cac9fae0ec7130a6b","success":1,"date_updated":"2023-12-04T08:04:14Z","file_size":57067476,"creator":"dernst","date_created":"2023-12-04T08:04:14Z","file_name":"2023_ACMToG_Makatura.pdf"}],"language":[{"iso":"eng"}],"issue":"5","volume":42,"abstract":[{"lang":"eng","text":"We introduce a compact, intuitive procedural graph representation for cellular metamaterials, which are small-scale, tileable structures that can be architected to exhibit many useful material properties. Because the structures’ “architectures” vary widely—with elements such as beams, thin shells, and solid bulks—it is difficult to explore them using existing representations. Generic approaches like voxel grids are versatile, but it is cumbersome to represent and edit individual structures; architecture-specific approaches address these issues, but are incompatible with one another. By contrast, our procedural graph succinctly represents the construction process for any structure using a simple skeleton annotated with spatially varying thickness. To express the highly constrained triply periodic minimal surfaces (TPMS) in this manner, we present the first fully automated version of the conjugate surface construction method, which allows novices to create complex TPMS from intuitive input. We demonstrate our representation’s expressiveness, accuracy, and compactness by constructing a wide range of established structures and hundreds of novel structures with diverse architectures and material properties. We also conduct a user study to verify our representation’s ease-of-use and ability to expand engineers’ capacity for exploration."}],"oa_version":"Published Version","month":"10","intvolume":" 42","citation":{"short":"L. Makatura, B. Wang, Y.-L. Chen, B. Deng, C. Wojtan, B. Bickel, W. Matusik, ACM Transactions on Graphics 42 (2023).","ieee":"L. Makatura et al., “Procedural metamaterials: A unified procedural graph for metamaterial design,” ACM Transactions on Graphics, vol. 42, no. 5. Association for Computing Machinery, 2023.","ama":"Makatura L, Wang B, Chen Y-L, et al. Procedural metamaterials: A unified procedural graph for metamaterial design. ACM Transactions on Graphics. 2023;42(5). doi:10.1145/3605389","apa":"Makatura, L., Wang, B., Chen, Y.-L., Deng, B., Wojtan, C., Bickel, B., & Matusik, W. (2023). Procedural metamaterials: A unified procedural graph for metamaterial design. ACM Transactions on Graphics. Association for Computing Machinery. https://doi.org/10.1145/3605389","mla":"Makatura, Liane, et al. “Procedural Metamaterials: A Unified Procedural Graph for Metamaterial Design.” ACM Transactions on Graphics, vol. 42, no. 5, 168, Association for Computing Machinery, 2023, doi:10.1145/3605389.","ista":"Makatura L, Wang B, Chen Y-L, Deng B, Wojtan C, Bickel B, Matusik W. 2023. Procedural metamaterials: A unified procedural graph for metamaterial design. ACM Transactions on Graphics. 42(5), 168.","chicago":"Makatura, Liane, Bohan Wang, Yi-Lu Chen, Bolei Deng, Chris Wojtan, Bernd Bickel, and Wojciech Matusik. “Procedural Metamaterials: A Unified Procedural Graph for Metamaterial Design.” ACM Transactions on Graphics. Association for Computing Machinery, 2023. https://doi.org/10.1145/3605389."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","author":[{"full_name":"Makatura, Liane","last_name":"Makatura","first_name":"Liane"},{"first_name":"Bohan","full_name":"Wang, Bohan","last_name":"Wang"},{"id":"0b467602-dbcd-11ea-9d1d-ed480aa46b70","first_name":"Yi-Lu","last_name":"Chen","full_name":"Chen, Yi-Lu"},{"first_name":"Bolei","last_name":"Deng","full_name":"Deng, Bolei"},{"id":"3C61F1D2-F248-11E8-B48F-1D18A9856A87","first_name":"Christopher J","orcid":"0000-0001-6646-5546","full_name":"Wojtan, Christopher J","last_name":"Wojtan"},{"last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd"},{"first_name":"Wojciech","last_name":"Matusik","full_name":"Matusik, Wojciech"}],"article_processing_charge":"Yes (in subscription journal)","title":"Procedural metamaterials: A unified procedural graph for metamaterial design","article_number":"168","project":[{"_id":"34bc2376-11ca-11ed-8bc3-9a3b3961a088","grant_number":"101045083","name":"Computational Discovery of Numerical Algorithms for Animation and Simulation of Natural Phenomena"}],"has_accepted_license":"1","year":"2023","day":"01","publication":"ACM Transactions on Graphics","date_published":"2023-10-01T00:00:00Z","doi":"10.1145/3605389","date_created":"2023-11-29T15:02:03Z","acknowledgement":"The authors thank Mina Konaković Luković and Michael Foshey for their early contributions to this project, David Palmer and Paul Zhang for their insightful discussions about minimal surfaces and the CSCM, Julian Panetta for providing the Elastic Textures code, and Hannes Hergeth for his feedback and support. We also thank our user study participants and anonymous reviewers.\r\nThis material is based upon work supported by the National Science Foundation\r\n(NSF) Graduate Research Fellowship under Grant No. 2141064; the MIT Morningside\r\nAcademy for Design Fellowship; the Defense Advanced Research Projects Agency\r\n(DARPA) Grant No. FA8750-20-C-0075; the ERC Consolidator Grant No. 101045083,\r\n“CoDiNA: Computational Discovery of Numerical Algorithms for Animation and Simulation of Natural Phenomena”; and the NewSat project, which is co-funded by the Operational Program for Competitiveness and Internationalisation (COMPETE2020), Portugal 2020, the European Regional Development Fund (ERDF), and the Portuguese Foundation for Science and Technology (FTC) under the MIT Portugal program.","publisher":"Association for Computing Machinery","quality_controlled":"1","oa":1},{"ddc":["004"],"date_updated":"2023-12-13T11:20:00Z","file_date_updated":"2023-05-16T09:12:05Z","department":[{"_id":"BeBi"}],"_id":"12976","keyword":["reinforcement learning","deposition","control","color","multi-filament"],"status":"public","conference":{"name":"ICRA: International Conference on Robotics and Automation","end_date":"2023-06-02","location":"London, United Kingdom","start_date":"2023-05-29"},"type":"conference","language":[{"iso":"eng"}],"file":[{"date_updated":"2023-05-16T09:12:05Z","file_size":5367986,"creator":"mpiovarc","date_created":"2023-05-16T09:12:05Z","file_name":"Liao2023.pdf","content_type":"application/pdf","access_level":"open_access","relation":"main_file","file_id":"12977","checksum":"daeaa67124777d88487f933ea3f77164","success":1}],"publication_status":"published","publication_identifier":{"issn":["1050-4729"],"eisbn":["9798350323658"]},"volume":2023,"oa_version":"Submitted Version","abstract":[{"lang":"eng","text":"3D printing based on continuous deposition of materials, such as filament-based 3D printing, has seen widespread adoption thanks to its versatility in working with a wide range of materials. An important shortcoming of this type of technology is its limited multi-material capabilities. While there are simple hardware designs that enable multi-material printing in principle, the required software is heavily underdeveloped. A typical hardware design fuses together individual materials fed into a single chamber from multiple inlets before they are deposited. This design, however, introduces a time delay between the intended material mixture and its actual deposition. In this work, inspired by diverse path planning research in robotics, we show that this mechanical challenge can be addressed via improved printer control. We propose to formulate the search for optimal multi-material printing policies in a reinforcement\r\nlearning setup. We put forward a simple numerical deposition model that takes into account the non-linear material mixing and delayed material deposition. To validate our system we focus on color fabrication, a problem known for its strict requirements for varying material mixtures at a high spatial frequency. We demonstrate that our learned control policy outperforms state-of-the-art hand-crafted algorithms."}],"intvolume":" 2023","month":"07","scopus_import":"1","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"ista":"Liao K, Tricard T, Piovarci M, Seidel H-P, Babaei V. 2023. Learning deposition policies for fused multi-material 3D printing. 2023 IEEE International Conference on Robotics and Automation. ICRA: International Conference on Robotics and Automation vol. 2023, 12345–12352.","chicago":"Liao, Kang, Thibault Tricard, Michael Piovarci, Hans-Peter Seidel, and Vahid Babaei. “Learning Deposition Policies for Fused Multi-Material 3D Printing.” In 2023 IEEE International Conference on Robotics and Automation, 2023:12345–52. IEEE, 2023. https://doi.org/10.1109/ICRA48891.2023.10160465.","short":"K. Liao, T. Tricard, M. Piovarci, H.-P. Seidel, V. Babaei, in:, 2023 IEEE International Conference on Robotics and Automation, IEEE, 2023, pp. 12345–12352.","ieee":"K. Liao, T. Tricard, M. Piovarci, H.-P. Seidel, and V. Babaei, “Learning deposition policies for fused multi-material 3D printing,” in 2023 IEEE International Conference on Robotics and Automation, London, United Kingdom, 2023, vol. 2023, pp. 12345–12352.","apa":"Liao, K., Tricard, T., Piovarci, M., Seidel, H.-P., & Babaei, V. (2023). Learning deposition policies for fused multi-material 3D printing. In 2023 IEEE International Conference on Robotics and Automation (Vol. 2023, pp. 12345–12352). London, United Kingdom: IEEE. https://doi.org/10.1109/ICRA48891.2023.10160465","ama":"Liao K, Tricard T, Piovarci M, Seidel H-P, Babaei V. Learning deposition policies for fused multi-material 3D printing. In: 2023 IEEE International Conference on Robotics and Automation. Vol 2023. IEEE; 2023:12345-12352. doi:10.1109/ICRA48891.2023.10160465","mla":"Liao, Kang, et al. “Learning Deposition Policies for Fused Multi-Material 3D Printing.” 2023 IEEE International Conference on Robotics and Automation, vol. 2023, IEEE, 2023, pp. 12345–52, doi:10.1109/ICRA48891.2023.10160465."},"title":"Learning deposition policies for fused multi-material 3D printing","article_processing_charge":"No","external_id":{"isi":["001048371104068"]},"author":[{"last_name":"Liao","full_name":"Liao, Kang","first_name":"Kang"},{"full_name":"Tricard, Thibault","last_name":"Tricard","first_name":"Thibault"},{"first_name":"Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","last_name":"Piovarci","full_name":"Piovarci, Michael","orcid":"0000-0002-5062-4474"},{"last_name":"Seidel","full_name":"Seidel, Hans-Peter","first_name":"Hans-Peter"},{"first_name":"Vahid","last_name":"Babaei","full_name":"Babaei, Vahid"}],"project":[{"_id":"eb901961-77a9-11ec-83b8-f5c883a62027","grant_number":"M03319","name":"Perception-Aware Appearance Fabrication"}],"publication":"2023 IEEE International Conference on Robotics and Automation","day":"04","year":"2023","has_accepted_license":"1","isi":1,"date_created":"2023-05-16T09:14:09Z","date_published":"2023-07-04T00:00:00Z","doi":"10.1109/ICRA48891.2023.10160465","page":"12345-12352","acknowledgement":"This work is graciously supported by FWF Lise Meitner (Grant M 3319). Kang Liao sincerely thank Emiliano Luci, Chunyu Lin, and Yao Zhao for their huge support.","oa":1,"quality_controlled":"1","publisher":"IEEE"},{"external_id":{"isi":["001018739600002"],"arxiv":["2201.02374"]},"article_processing_charge":"No","author":[{"last_name":"Zhong","full_name":"Zhong, Fanchao","first_name":"Fanchao"},{"full_name":"Xu, Yonglai","last_name":"Xu","first_name":"Yonglai"},{"orcid":"0000-0002-6389-1045","full_name":"Zhao, Haisen","last_name":"Zhao","id":"fb7f793a-80d1-11eb-8869-d56e5b2a8ff4","first_name":"Haisen"},{"full_name":"Lu, Lin","last_name":"Lu","first_name":"Lin"}],"title":"As-Continuous-As-Possible extrusion-based fabrication of surface models","citation":{"ista":"Zhong F, Xu Y, Zhao H, Lu L. 2023. As-Continuous-As-Possible extrusion-based fabrication of surface models. ACM Transactions on Graphics. 42(3), 26.","chicago":"Zhong, Fanchao, Yonglai Xu, Haisen Zhao, and Lin Lu. “As-Continuous-As-Possible Extrusion-Based Fabrication of Surface Models.” ACM Transactions on Graphics. Association for Computing Machinery, 2023. https://doi.org/10.1145/3575859.","ieee":"F. Zhong, Y. Xu, H. Zhao, and L. Lu, “As-Continuous-As-Possible extrusion-based fabrication of surface models,” ACM Transactions on Graphics, vol. 42, no. 3. Association for Computing Machinery, 2023.","short":"F. Zhong, Y. Xu, H. Zhao, L. Lu, ACM Transactions on Graphics 42 (2023).","ama":"Zhong F, Xu Y, Zhao H, Lu L. As-Continuous-As-Possible extrusion-based fabrication of surface models. ACM Transactions on Graphics. 2023;42(3). doi:10.1145/3575859","apa":"Zhong, F., Xu, Y., Zhao, H., & Lu, L. (2023). As-Continuous-As-Possible extrusion-based fabrication of surface models. ACM Transactions on Graphics. Association for Computing Machinery. https://doi.org/10.1145/3575859","mla":"Zhong, Fanchao, et al. “As-Continuous-As-Possible Extrusion-Based Fabrication of Surface Models.” ACM Transactions on Graphics, vol. 42, no. 3, 26, Association for Computing Machinery, 2023, doi:10.1145/3575859."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_number":"26","date_created":"2023-07-23T22:01:13Z","date_published":"2023-03-17T00:00:00Z","doi":"10.1145/3575859","year":"2023","isi":1,"publication":"ACM Transactions on Graphics","day":"17","oa":1,"quality_controlled":"1","publisher":"Association for Computing Machinery","acknowledgement":"This work was supported in part by grants from the NSFC (61972232), Science and Technology Program of Shenzhen, China (CJGJZD20200617102202007). ","department":[{"_id":"BeBi"}],"date_updated":"2023-12-13T11:34:59Z","type":"journal_article","article_type":"original","status":"public","_id":"13265","issue":"3","volume":42,"publication_status":"published","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"language":[{"iso":"eng"}],"main_file_link":[{"open_access":"1","url":"https://doi.org/10.48550/arXiv.2201.02374"}],"scopus_import":"1","intvolume":" 42","month":"03","abstract":[{"text":"In this study, we propose a computational framework for optimizing the continuity of the toolpath in fabricating surface models on an extrusion-based 3D printer. Toolpath continuity is a critical issue that influences both the quality and the efficiency of extrusion-based fabrication. Transfer moves lead to rough and bumpy surfaces, where this phenomenon worsens for materials with large viscosity, like clay. The effects of continuity on the surface models are even more severe in terms of the quality of the surface and the stability of the model. We introduce a criterion called the one–path patch (OPP) to represent a patch on the surface of the shell that can be traversed along one path by considering the constraints on fabrication. We study the properties of the OPPs and their merging operations to propose a bottom-up OPP merging procedure to decompose the given shell surface into a minimal number of OPPs, and to generate the “as-continuous-as-possible” (ACAP) toolpath. Furthermore, we augment the path planning algorithm with a curved-layer printing scheme that reduces staircase defects and improves the continuity of the toolpath by connecting multiple segments. We evaluated the ACAP algorithm on ceramic and thermoplastic materials, and the results showed that it improves the fabrication of surface models in terms of both efficiency and surface quality.","lang":"eng"}],"oa_version":"Preprint"},{"ec_funded":1,"volume":20,"related_material":{"record":[{"relation":"research_data","status":"public","id":"12817"},{"relation":"shorter_version","status":"public","id":"14770"}],"link":[{"url":"https://github.com/danzllab/LIONESS","relation":"software"}]},"publication_status":"published","publication_identifier":{"issn":["1548-7091"],"eissn":["1548-7105"]},"language":[{"iso":"eng"}],"main_file_link":[{"open_access":"1","url":"https://doi.org/10.1038/s41592-023-01936-6"}],"scopus_import":"1","intvolume":" 20","month":"08","acknowledged_ssus":[{"_id":"ScienComp"},{"_id":"Bio"},{"_id":"PreCl"},{"_id":"E-Lib"},{"_id":"LifeSc"},{"_id":"M-Shop"}],"abstract":[{"lang":"eng","text":"Three-dimensional (3D) reconstruction of living brain tissue down to an individual synapse level would create opportunities for decoding the dynamics and structure–function relationships of the brain’s complex and dense information processing network; however, this has been hindered by insufficient 3D resolution, inadequate signal-to-noise ratio and prohibitive light burden in optical imaging, whereas electron microscopy is inherently static. Here we solved these challenges by developing an integrated optical/machine-learning technology, LIONESS (live information-optimized nanoscopy enabling saturated segmentation). This leverages optical modifications to stimulated emission depletion microscopy in comprehensively, extracellularly labeled tissue and previous information on sample structure via machine learning to simultaneously achieve isotropic super-resolution, high signal-to-noise ratio and compatibility with living tissue. This allows dense deep-learning-based instance segmentation and 3D reconstruction at a synapse level, incorporating molecular, activity and morphodynamic information. LIONESS opens up avenues for studying the dynamic functional (nano-)architecture of living brain tissue."}],"oa_version":"Published Version","pmid":1,"department":[{"_id":"PeJo"},{"_id":"GaNo"},{"_id":"BeBi"},{"_id":"JoDa"},{"_id":"Bio"}],"date_updated":"2024-01-10T08:37:48Z","article_type":"original","type":"journal_article","status":"public","_id":"13267","page":"1256-1265","date_created":"2023-07-23T22:01:13Z","doi":"10.1038/s41592-023-01936-6","date_published":"2023-08-01T00:00:00Z","year":"2023","isi":1,"publication":"Nature Methods","day":"01","oa":1,"publisher":"Springer Nature","quality_controlled":"1","acknowledgement":"We thank J. Vorlaufer, N. Agudelo and A. Wartak for microscope maintenance and troubleshooting, C. Kreuzinger and A. Freeman for technical assistance, M. Šuplata for hardware control support and M. Cunha dos Santos for initial exploration of software. We\r\nthank P. Henderson for advice on deep-learning training and M. Sixt, S. Boyd and T. Weiss for discussions and critical reading of the manuscript. L. Lavis (Janelia Research Campus) generously provided the JF585-HaloTag ligand. We acknowledge expert support by IST\r\nAustria’s scientific computing, imaging and optics, preclinical, library and laboratory support facilities and by the Miba machine shop. We gratefully acknowledge funding by the following sources: Austrian Science Fund (F.W.F.) grant no. I3600-B27 (J.G.D.), grant no. DK W1232\r\n(J.G.D. and J.M.M.) and grant no. Z 312-B27, Wittgenstein award (P.J.); the Gesellschaft für Forschungsförderung NÖ grant no. LSC18-022 (J.G.D.); an ISTA Interdisciplinary project grant (J.G.D. and B.B.); the European Union’s Horizon 2020 research and innovation programme,\r\nMarie-Skłodowska Curie grant 665385 (J.M.M. and J.L.); the European Union’s Horizon 2020 research and innovation programme, European Research Council grant no. 715767, MATERIALIZABLE (B.B.); grant no. 715508, REVERSEAUTISM (G.N.); grant no. 695568, SYNNOVATE (S.G.N.G.); and grant no. 692692, GIANTSYN (P.J.); the Simons\r\nFoundation Autism Research Initiative grant no. 529085 (S.G.N.G.); the Wellcome Trust Technology Development grant no. 202932 (S.G.N.G.); the Marie Skłodowska-Curie Actions Individual Fellowship no. 101026635 under the EU Horizon 2020 program (J.F.W.);\r\nthe Human Frontier Science Program postdoctoral fellowship LT000557/2018 (W.J.); and the National Science Foundation grant no. IIS-1835231 (H.P.) and NCS-FO-2124179 (H.P.).","external_id":{"isi":["001025621500001"],"pmid":["37429995"]},"article_processing_charge":"Yes","author":[{"full_name":"Velicky, Philipp","orcid":"0000-0002-2340-7431","last_name":"Velicky","id":"39BDC62C-F248-11E8-B48F-1D18A9856A87","first_name":"Philipp"},{"id":"3FB91342-F248-11E8-B48F-1D18A9856A87","first_name":"Eder","last_name":"Miguel Villalba","orcid":"0000-0001-5665-0430","full_name":"Miguel Villalba, Eder"},{"first_name":"Julia M","id":"443DB6DE-F248-11E8-B48F-1D18A9856A87","last_name":"Michalska","full_name":"Michalska, Julia M","orcid":"0000-0003-3862-1235"},{"first_name":"Julia","id":"46E28B80-F248-11E8-B48F-1D18A9856A87","full_name":"Lyudchik, Julia","last_name":"Lyudchik"},{"first_name":"Donglai","last_name":"Wei","full_name":"Wei, Donglai"},{"last_name":"Lin","full_name":"Lin, Zudi","first_name":"Zudi"},{"last_name":"Watson","orcid":"0000-0002-8698-3823","full_name":"Watson, Jake","id":"63836096-4690-11EA-BD4E-32803DDC885E","first_name":"Jake"},{"full_name":"Troidl, Jakob","last_name":"Troidl","first_name":"Jakob"},{"full_name":"Beyer, Johanna","last_name":"Beyer","first_name":"Johanna"},{"id":"43DF3136-F248-11E8-B48F-1D18A9856A87","first_name":"Yoav","full_name":"Ben Simon, Yoav","last_name":"Ben Simon"},{"last_name":"Sommer","full_name":"Sommer, Christoph M","orcid":"0000-0003-1216-9105","first_name":"Christoph M","id":"4DF26D8C-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Wiebke","id":"425C1CE8-F248-11E8-B48F-1D18A9856A87","full_name":"Jahr, Wiebke","last_name":"Jahr"},{"full_name":"Cenameri, Alban","last_name":"Cenameri","first_name":"Alban","id":"9ac8f577-2357-11eb-997a-e566c5550886"},{"first_name":"Johannes","full_name":"Broichhagen, Johannes","last_name":"Broichhagen"},{"full_name":"Grant, Seth G.N.","last_name":"Grant","first_name":"Seth G.N."},{"id":"353C1B58-F248-11E8-B48F-1D18A9856A87","first_name":"Peter M","last_name":"Jonas","full_name":"Jonas, Peter M","orcid":"0000-0001-5001-4804"},{"last_name":"Novarino","orcid":"0000-0002-7673-7178","full_name":"Novarino, Gaia","id":"3E57A680-F248-11E8-B48F-1D18A9856A87","first_name":"Gaia"},{"full_name":"Pfister, Hanspeter","last_name":"Pfister","first_name":"Hanspeter"},{"last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd"},{"id":"42EFD3B6-F248-11E8-B48F-1D18A9856A87","first_name":"Johann G","full_name":"Danzl, Johann G","orcid":"0000-0001-8559-3973","last_name":"Danzl"}],"title":"Dense 4D nanoscale reconstruction of living brain tissue","citation":{"ista":"Velicky P, Miguel Villalba E, Michalska JM, Lyudchik J, Wei D, Lin Z, Watson J, Troidl J, Beyer J, Ben Simon Y, Sommer CM, Jahr W, Cenameri A, Broichhagen J, Grant SGN, Jonas PM, Novarino G, Pfister H, Bickel B, Danzl JG. 2023. Dense 4D nanoscale reconstruction of living brain tissue. Nature Methods. 20, 1256–1265.","chicago":"Velicky, Philipp, Eder Miguel Villalba, Julia M Michalska, Julia Lyudchik, Donglai Wei, Zudi Lin, Jake Watson, et al. “Dense 4D Nanoscale Reconstruction of Living Brain Tissue.” Nature Methods. Springer Nature, 2023. https://doi.org/10.1038/s41592-023-01936-6.","ama":"Velicky P, Miguel Villalba E, Michalska JM, et al. Dense 4D nanoscale reconstruction of living brain tissue. Nature Methods. 2023;20:1256-1265. doi:10.1038/s41592-023-01936-6","apa":"Velicky, P., Miguel Villalba, E., Michalska, J. M., Lyudchik, J., Wei, D., Lin, Z., … Danzl, J. G. (2023). Dense 4D nanoscale reconstruction of living brain tissue. Nature Methods. Springer Nature. https://doi.org/10.1038/s41592-023-01936-6","short":"P. Velicky, E. Miguel Villalba, J.M. Michalska, J. Lyudchik, D. Wei, Z. Lin, J. Watson, J. Troidl, J. Beyer, Y. Ben Simon, C.M. Sommer, W. Jahr, A. Cenameri, J. Broichhagen, S.G.N. Grant, P.M. Jonas, G. Novarino, H. Pfister, B. Bickel, J.G. Danzl, Nature Methods 20 (2023) 1256–1265.","ieee":"P. Velicky et al., “Dense 4D nanoscale reconstruction of living brain tissue,” Nature Methods, vol. 20. Springer Nature, pp. 1256–1265, 2023.","mla":"Velicky, Philipp, et al. “Dense 4D Nanoscale Reconstruction of Living Brain Tissue.” Nature Methods, vol. 20, Springer Nature, 2023, pp. 1256–65, doi:10.1038/s41592-023-01936-6."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","project":[{"grant_number":"I03600","name":"Optical control of synaptic function via adhesion molecules","_id":"265CB4D0-B435-11E9-9278-68D0E5697425","call_identifier":"FWF"},{"call_identifier":"FWF","_id":"2548AE96-B435-11E9-9278-68D0E5697425","name":"Molecular Drug Targets","grant_number":"W1232-B24"},{"_id":"25C5A090-B435-11E9-9278-68D0E5697425","call_identifier":"FWF","name":"The Wittgenstein Prize","grant_number":"Z00312"},{"_id":"23889792-32DE-11EA-91FC-C7463DDC885E","name":"High content imaging to decode human immune cell interactions in health and allergic disease"},{"name":"International IST Doctoral Program","grant_number":"665385","call_identifier":"H2020","_id":"2564DBCA-B435-11E9-9278-68D0E5697425"},{"call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767"},{"_id":"25444568-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","grant_number":"715508","name":"Probing the Reversibility of Autism Spectrum Disorders by Employing in vivo and in vitro Models"},{"_id":"25B7EB9E-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"Biophysics and circuit function of a giant cortical glumatergic synapse","grant_number":"692692"},{"name":"Synaptic computations of the hippocampal CA3 circuitry","grant_number":"101026635","_id":"fc2be41b-9c52-11eb-aca3-faa90aa144e9","call_identifier":"H2020"},{"_id":"2668BFA0-B435-11E9-9278-68D0E5697425","grant_number":"LT00057","name":"High-speed 3D-nanoscopy to study the role of adhesion during 3D cell migration"}]},{"citation":{"ista":"Chen B, Jindal A, Piovarci M, Wang C, Seidel HP, Didyk P, Myszkowski K, Serrano A, Mantiuk RK. 2023. The effect of display capabilities on the gloss consistency between real and virtual objects. Proceedings of the SIGGRAPH Asia 2023 Conference. SIGGRAPH: Computer Graphics and Interactive Techniques Conference, 90.","chicago":"Chen, Bin, Akshay Jindal, Michael Piovarci, Chao Wang, Hans Peter Seidel, Piotr Didyk, Karol Myszkowski, Ana Serrano, and Rafał K. Mantiuk. “The Effect of Display Capabilities on the Gloss Consistency between Real and Virtual Objects.” In Proceedings of the SIGGRAPH Asia 2023 Conference. Association for Computing Machinery, 2023. https://doi.org/10.1145/3610548.3618226.","ama":"Chen B, Jindal A, Piovarci M, et al. The effect of display capabilities on the gloss consistency between real and virtual objects. In: Proceedings of the SIGGRAPH Asia 2023 Conference. Association for Computing Machinery; 2023. doi:10.1145/3610548.3618226","apa":"Chen, B., Jindal, A., Piovarci, M., Wang, C., Seidel, H. P., Didyk, P., … Mantiuk, R. K. (2023). The effect of display capabilities on the gloss consistency between real and virtual objects. In Proceedings of the SIGGRAPH Asia 2023 Conference. Sydney, Australia: Association for Computing Machinery. https://doi.org/10.1145/3610548.3618226","ieee":"B. Chen et al., “The effect of display capabilities on the gloss consistency between real and virtual objects,” in Proceedings of the SIGGRAPH Asia 2023 Conference, Sydney, Australia, 2023.","short":"B. Chen, A. Jindal, M. Piovarci, C. Wang, H.P. Seidel, P. Didyk, K. Myszkowski, A. Serrano, R.K. Mantiuk, in:, Proceedings of the SIGGRAPH Asia 2023 Conference, Association for Computing Machinery, 2023.","mla":"Chen, Bin, et al. “The Effect of Display Capabilities on the Gloss Consistency between Real and Virtual Objects.” Proceedings of the SIGGRAPH Asia 2023 Conference, 90, Association for Computing Machinery, 2023, doi:10.1145/3610548.3618226."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","author":[{"first_name":"Bin","full_name":"Chen, Bin","last_name":"Chen"},{"full_name":"Jindal, Akshay","last_name":"Jindal","first_name":"Akshay"},{"full_name":"Piovarci, Michael","orcid":"0000-0002-5062-4474","last_name":"Piovarci","first_name":"Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E"},{"last_name":"Wang","full_name":"Wang, Chao","first_name":"Chao"},{"full_name":"Seidel, Hans Peter","last_name":"Seidel","first_name":"Hans Peter"},{"last_name":"Didyk","full_name":"Didyk, Piotr","first_name":"Piotr"},{"first_name":"Karol","last_name":"Myszkowski","full_name":"Myszkowski, Karol"},{"first_name":"Ana","last_name":"Serrano","full_name":"Serrano, Ana"},{"full_name":"Mantiuk, Rafał K.","last_name":"Mantiuk","first_name":"Rafał K."}],"article_processing_charge":"Yes (in subscription journal)","title":"The effect of display capabilities on the gloss consistency between real and virtual objects","article_number":"90","project":[{"_id":"eb901961-77a9-11ec-83b8-f5c883a62027","name":"Perception-Aware Appearance Fabrication","grant_number":"M03319"}],"has_accepted_license":"1","year":"2023","day":"10","publication":"Proceedings of the SIGGRAPH Asia 2023 Conference","doi":"10.1145/3610548.3618226","date_published":"2023-12-10T00:00:00Z","date_created":"2024-01-14T23:00:57Z","acknowledgement":"This work is supported by FWF Lise Meitner (Grant M 3319), Spanish Agencia Estatal de Investigación (project PID2022-141539NBI00), European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement\r\nN◦ 725253–EyeCode), Swiss National Science Foundation (Grant no. 200502), and academic gifts from Meta. We thank Dmitry Lubyako and Ali Özgür Yöntem for building the turntable for our experiment.","quality_controlled":"1","publisher":"Association for Computing Machinery","oa":1,"date_updated":"2024-01-17T08:38:35Z","ddc":["000"],"file_date_updated":"2024-01-17T08:33:06Z","department":[{"_id":"BeBi"}],"_id":"14798","type":"conference","tmp":{"legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","short":"CC BY (4.0)"},"conference":{"name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference","start_date":"2023-12-12","end_date":"2023-12-15","location":"Sydney, Australia"},"status":"public","publication_identifier":{"isbn":["9798400703157"]},"publication_status":"published","file":[{"access_level":"open_access","relation":"main_file","content_type":"application/pdf","file_id":"14823","checksum":"8abe27432ed222b50d1af9b3388db1b0","success":1,"creator":"dernst","date_updated":"2024-01-17T08:33:06Z","file_size":95967451,"date_created":"2024-01-17T08:33:06Z","file_name":"2023_SA_Chen.pdf"}],"language":[{"iso":"eng"}],"abstract":[{"lang":"eng","text":"A faithful reproduction of gloss is inherently difficult because of the limited dynamic range, peak luminance, and 3D capabilities of display devices. This work investigates how the display capabilities affect gloss appearance with respect to a real-world reference object. To this end, we employ an accurate imaging pipeline to achieve a perceptual gloss match between a virtual and real object presented side-by-side on an augmented-reality high-dynamic-range (HDR) stereoscopic display, which has not been previously attained to this extent. Based on this precise gloss reproduction, we conduct a series of gloss matching experiments to study how gloss perception degrades based on individual factors: object albedo, display luminance, dynamic range, stereopsis, and tone mapping. We support the study with a detailed analysis of individual factors, followed by an in-depth discussion on the observed perceptual effects. Our experiments demonstrate that stereoscopic presentation has a limited effect on the gloss matching task on our HDR display. However, both reduced luminance and dynamic range of the display reduce the perceived gloss. This means that the visual system cannot compensate for the changes in gloss appearance across luminance (lack of gloss constancy), and the tone mapping operator should be carefully selected when reproducing gloss on a low dynamic range (LDR) display."}],"oa_version":"Published Version","scopus_import":"1","month":"12"},{"_id":"13049","keyword":["PCB design and layout","Mesh geometry models"],"status":"public","conference":{"start_date":"2023-08-06","location":"Los Angeles, CA, United States","end_date":"2023-08-10","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference"},"article_type":"original","type":"journal_article","ddc":["006"],"date_updated":"2024-01-29T10:30:49Z","department":[{"_id":"GradSch"},{"_id":"BeBi"}],"file_date_updated":"2023-06-20T12:20:51Z","oa_version":"Submitted Version","acknowledged_ssus":[{"_id":"M-Shop"}],"abstract":[{"text":"We propose a computational design approach for covering a surface with individually addressable RGB LEDs, effectively forming a low-resolution surface screen. To achieve a low-cost and scalable approach, we propose creating designs from flat PCB panels bent in-place along the surface of a 3D printed core. Working with standard rigid PCBs enables the use of\r\nestablished PCB manufacturing services, allowing the fabrication of designs with several hundred LEDs. \r\nOur approach optimizes the PCB geometry for folding, and then jointly optimizes the LED packing, circuit and routing, solving a challenging layout problem under strict manufacturing requirements. Unlike paper, PCBs cannot bend beyond a certain point without breaking. Therefore, we introduce parametric cut patterns acting as hinges, designed to allow bending while remaining compact. To tackle the joint optimization of placement, circuit and routing, we propose a specialized algorithm that splits the global problem into one sub-problem per triangle, which is then individually solved.\r\nOur technique generates PCB blueprints in a completely automated way. After being fabricated by a PCB manufacturing service, the boards are bent and glued by the user onto the 3D printed support. We demonstrate our technique on a range of physical models and virtual examples, creating intricate surface light patterns from hundreds of LEDs.","lang":"eng"}],"intvolume":" 42","month":"07","language":[{"iso":"eng"}],"file":[{"success":1,"file_id":"13156","checksum":"a0b0ba3b36f43a94388e8824613d812a","relation":"main_file","access_level":"open_access","content_type":"application/pdf","file_name":"2023_ACMToG_Freire.pdf","date_created":"2023-06-19T11:02:23Z","creator":"dernst","file_size":78940724,"date_updated":"2023-06-19T11:02:23Z"},{"checksum":"b9206bbb67af82df49b7e7cdbde3410c","file_id":"13157","success":1,"access_level":"open_access","relation":"main_file","content_type":"application/pdf","date_created":"2023-06-20T12:20:51Z","file_name":"2023_ACMToG_SuppMaterial_Freire.pdf","creator":"dernst","date_updated":"2023-06-20T12:20:51Z","file_size":34345905}],"publication_status":"published","publication_identifier":{"issn":["0730-0301"],"eissn":["1557-7368"]},"ec_funded":1,"issue":"4","volume":42,"article_number":"142","project":[{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"}],"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"ista":"Freire M, Bhargava M, Schreck C, Hugron P-A, Bickel B, Lefebvre S. 2023. PCBend: Light up your 3D shapes with foldable circuit boards. Transactions on Graphics. 42(4), 142.","chicago":"Freire, Marco, Manas Bhargava, Camille Schreck, Pierre-Alexandre Hugron, Bernd Bickel, and Sylvain Lefebvre. “PCBend: Light up Your 3D Shapes with Foldable Circuit Boards.” Transactions on Graphics. Association for Computing Machinery, 2023. https://doi.org/10.1145/3592411.","apa":"Freire, M., Bhargava, M., Schreck, C., Hugron, P.-A., Bickel, B., & Lefebvre, S. (2023). PCBend: Light up your 3D shapes with foldable circuit boards. Transactions on Graphics. Los Angeles, CA, United States: Association for Computing Machinery. https://doi.org/10.1145/3592411","ama":"Freire M, Bhargava M, Schreck C, Hugron P-A, Bickel B, Lefebvre S. PCBend: Light up your 3D shapes with foldable circuit boards. Transactions on Graphics. 2023;42(4). doi:10.1145/3592411","short":"M. Freire, M. Bhargava, C. Schreck, P.-A. Hugron, B. Bickel, S. Lefebvre, Transactions on Graphics 42 (2023).","ieee":"M. Freire, M. Bhargava, C. Schreck, P.-A. Hugron, B. Bickel, and S. Lefebvre, “PCBend: Light up your 3D shapes with foldable circuit boards,” Transactions on Graphics, vol. 42, no. 4. Association for Computing Machinery, 2023.","mla":"Freire, Marco, et al. “PCBend: Light up Your 3D Shapes with Foldable Circuit Boards.” Transactions on Graphics, vol. 42, no. 4, 142, Association for Computing Machinery, 2023, doi:10.1145/3592411."},"title":"PCBend: Light up your 3D shapes with foldable circuit boards","article_processing_charge":"No","external_id":{"isi":["001044671300108"]},"author":[{"full_name":"Freire, Marco","last_name":"Freire","first_name":"Marco"},{"id":"FF8FA64C-AA6A-11E9-99AD-50D4E5697425","first_name":"Manas","last_name":"Bhargava","full_name":"Bhargava, Manas","orcid":"0009-0007-6138-6890"},{"full_name":"Schreck, Camille","last_name":"Schreck","id":"2B14B676-F248-11E8-B48F-1D18A9856A87","first_name":"Camille"},{"full_name":"Hugron, Pierre-Alexandre","last_name":"Hugron","first_name":"Pierre-Alexandre"},{"last_name":"Bickel","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd","first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Sylvain","last_name":"Lefebvre","full_name":"Lefebvre, Sylvain"}],"acknowledgement":"We thank the reviewers for the valuable feedback. We also thank the Miba Machine Shop at ISTA, PCBWay, and PragoBoard for helping us with fabrication and assembly. This project was supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (Grant Agreement No. 715767 – MATERIALIZABLE).","oa":1,"quality_controlled":"1","publisher":"Association for Computing Machinery","publication":"Transactions on Graphics","day":"26","year":"2023","isi":1,"has_accepted_license":"1","date_created":"2023-05-22T08:37:04Z","date_published":"2023-07-26T00:00:00Z","doi":"10.1145/3592411"},{"year":"2023","has_accepted_license":"1","isi":1,"publication":"Transactions on Graphics","day":"26","date_created":"2023-05-16T09:39:14Z","date_published":"2023-07-26T00:00:00Z","doi":"10.1145/3592432","acknowledgement":"We thank Todor Asenov and the Miba Machine Shop for their help in assembling the tattoo machine and manufacturing the substrates. We thank Geysler Rodrigues for the insightful discussions on tattooing practices from a professional artist's perspective. We thank Maria Fernanda Portugal for sharing a doctor's perspective on medical applications of tattoos. This work is graciously supported by the FWF Lise Meitner (Grant M 3319).","oa":1,"publisher":"Association for Computing Machinery","quality_controlled":"1","citation":{"mla":"Piovarci, Michael, et al. “Skin-Screen: A Computational Fabrication Framework for Color Tattoos.” Transactions on Graphics, vol. 42, no. 4, 67, Association for Computing Machinery, 2023, doi:10.1145/3592432.","short":"M. Piovarci, A. Chapiro, B. Bickel, Transactions on Graphics 42 (2023).","ieee":"M. Piovarci, A. Chapiro, and B. Bickel, “Skin-Screen: A computational fabrication framework for color tattoos,” Transactions on Graphics, vol. 42, no. 4. Association for Computing Machinery, 2023.","apa":"Piovarci, M., Chapiro, A., & Bickel, B. (2023). Skin-Screen: A computational fabrication framework for color tattoos. Transactions on Graphics. Los Angeles, CA, United States: Association for Computing Machinery. https://doi.org/10.1145/3592432","ama":"Piovarci M, Chapiro A, Bickel B. Skin-Screen: A computational fabrication framework for color tattoos. Transactions on Graphics. 2023;42(4). doi:10.1145/3592432","chicago":"Piovarci, Michael, Alexandre Chapiro, and Bernd Bickel. “Skin-Screen: A Computational Fabrication Framework for Color Tattoos.” Transactions on Graphics. Association for Computing Machinery, 2023. https://doi.org/10.1145/3592432.","ista":"Piovarci M, Chapiro A, Bickel B. 2023. Skin-Screen: A computational fabrication framework for color tattoos. Transactions on Graphics. 42(4), 67."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","external_id":{"isi":["001044671300033"]},"author":[{"last_name":"Piovarci","full_name":"Piovarci, Michael","orcid":"0000-0002-5062-4474","first_name":"Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E"},{"first_name":"Alexandre","last_name":"Chapiro","full_name":"Chapiro, Alexandre"},{"full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd"}],"title":"Skin-Screen: A computational fabrication framework for color tattoos","article_number":"67","project":[{"name":"Perception-Aware Appearance Fabrication","grant_number":"M03319","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"}],"publication_status":"published","publication_identifier":{"issn":["0730-0301"],"eissn":["1557-7368"]},"language":[{"iso":"eng"}],"file":[{"success":1,"checksum":"5f0a6867689e025a661bd0b4fd90b821","file_id":"12985","content_type":"application/pdf","relation":"main_file","access_level":"open_access","file_name":"Piovarci2023.pdf","date_created":"2023-05-16T09:38:25Z","file_size":30817343,"date_updated":"2023-05-16T09:38:25Z","creator":"mpiovarc"}],"volume":42,"issue":"4","abstract":[{"text":"Tattoos are a highly popular medium, with both artistic and medical applications. Although the mechanical process of tattoo application has evolved historically, the results are reliant on the artisanal skill of the artist. This can be especially challenging for some skin tones, or in cases where artists lack experience. We provide the first systematic overview of tattooing as a computational fabrication technique. We built an automated tattooing rig and a recipe for the creation of silicone sheets mimicking realistic skin tones, which allowed us to create an accurate model predicting tattoo appearance. This enables several exciting applications including tattoo previewing, color retargeting, novel ink spectra optimization, color-accurate prosthetics, and more.","lang":"eng"}],"acknowledged_ssus":[{"_id":"M-Shop"}],"oa_version":"Submitted Version","intvolume":" 42","month":"07","date_updated":"2024-01-29T10:27:23Z","ddc":["004"],"department":[{"_id":"BeBi"}],"file_date_updated":"2023-05-16T09:38:25Z","_id":"12984","conference":{"start_date":"2023-08-06","location":"Los Angeles, CA, United States","end_date":"2023-08-10","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference"},"article_type":"original","type":"journal_article","keyword":["appearance","modeling","reproduction","tattoo","skin color","gamut mapping","ink-optimization","prosthetic"],"status":"public"},{"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"ista":"Condor J, Piovarci M, Bickel B, Didyk P. 2023. Gloss-aware color correction for 3D printing. SIGGRAPH ’23 Conference Proceedings. SIGGRAPH: Computer Graphics and Interactive Techniques Conference, 21.","chicago":"Condor, Jorge, Michael Piovarci, Bernd Bickel, and Piotr Didyk. “Gloss-Aware Color Correction for 3D Printing.” In SIGGRAPH ’23 Conference Proceedings. Association for Computing Machinery, 2023. https://doi.org/10.1145/3588432.3591546.","short":"J. Condor, M. Piovarci, B. Bickel, P. Didyk, in:, SIGGRAPH ’23 Conference Proceedings, Association for Computing Machinery, 2023.","ieee":"J. Condor, M. Piovarci, B. Bickel, and P. Didyk, “Gloss-aware color correction for 3D printing,” in SIGGRAPH ’23 Conference Proceedings, Los Angeles, CA, United States, 2023.","apa":"Condor, J., Piovarci, M., Bickel, B., & Didyk, P. (2023). Gloss-aware color correction for 3D printing. In SIGGRAPH ’23 Conference Proceedings. Los Angeles, CA, United States: Association for Computing Machinery. https://doi.org/10.1145/3588432.3591546","ama":"Condor J, Piovarci M, Bickel B, Didyk P. Gloss-aware color correction for 3D printing. In: SIGGRAPH ’23 Conference Proceedings. Association for Computing Machinery; 2023. doi:10.1145/3588432.3591546","mla":"Condor, Jorge, et al. “Gloss-Aware Color Correction for 3D Printing.” SIGGRAPH ’23 Conference Proceedings, 21, Association for Computing Machinery, 2023, doi:10.1145/3588432.3591546."},"title":"Gloss-aware color correction for 3D printing","article_processing_charge":"No","external_id":{"isi":["001117690500021"]},"author":[{"full_name":"Condor, Jorge","last_name":"Condor","first_name":"Jorge"},{"orcid":"0000-0002-5062-4474","full_name":"Piovarci, Michael","last_name":"Piovarci","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael"},{"orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd","last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd"},{"first_name":"Piotr","full_name":"Didyk, Piotr","last_name":"Didyk"}],"article_number":"21","project":[{"name":"Perception-Aware Appearance Fabrication","grant_number":"M03319","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"}],"publication":"SIGGRAPH ’23 Conference Proceedings","day":"23","year":"2023","has_accepted_license":"1","isi":1,"date_created":"2023-05-16T09:34:13Z","doi":"10.1145/3588432.3591546","date_published":"2023-07-23T00:00:00Z","acknowledgement":"We thank Matthew S Zurawski for the 3D model of the car speed shape. This research has been supported by the Swiss National Science Foundation (SNSF, Grant 200502) and the FWF Lise Meitner (Grant M 3319).","oa":1,"quality_controlled":"1","publisher":"Association for Computing Machinery","ddc":["004"],"date_updated":"2024-02-28T12:52:04Z","department":[{"_id":"BeBi"}],"file_date_updated":"2024-01-29T10:14:10Z","_id":"12979","keyword":["color","gloss","perception","color compensation","color management"],"status":"public","conference":{"start_date":"2023-08-06","end_date":"2023-08-10","location":"Los Angeles, CA, United States","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference"},"tmp":{"legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","short":"CC BY (4.0)"},"type":"conference","language":[{"iso":"eng"}],"file":[{"content_type":"application/pdf","access_level":"open_access","relation":"main_file","checksum":"84a437739af5d46507928939b20c0c28","file_id":"12983","success":1,"date_updated":"2023-05-16T09:32:50Z","file_size":42323971,"creator":"mpiovarc","date_created":"2023-05-16T09:32:50Z","file_name":"Condor2023_supplemental.pdf"},{"content_type":"application/pdf","access_level":"open_access","relation":"main_file","checksum":"0f5c8b242e8e7c153c04888c4d0c6f37","file_id":"14893","success":1,"date_updated":"2024-01-29T10:14:10Z","file_size":26079404,"creator":"dernst","date_created":"2024-01-29T10:14:10Z","file_name":"2023_Siggraph_Condor.pdf"}],"publication_status":"published","publication_identifier":{"isbn":["9798400701597"]},"oa_version":"Published Version","abstract":[{"text":"Color and gloss are fundamental aspects of surface appearance. State-of-the-art fabrication techniques can manipulate both properties of the printed 3D objects. However, in the context of appearance reproduction, perceptual aspects of color and gloss are usually handled separately, even though previous perceptual studies suggest their interaction. Our work is motivated by previous studies demonstrating a perceived color shift due to a change in the object's gloss, i.e., two samples with the same color but different surface gloss appear as they have different colors. In this paper, we conduct new experiments which support this observation and provide insights into the magnitude and direction of the perceived color change. We use the observations as guidance to design a new method that estimates and corrects the color shift enabling the fabrication of objects with the same perceived color but different surface gloss. We formulate the problem as an optimization procedure solved using differentiable rendering. We evaluate the effectiveness of our method in perceptual experiments with 3D objects fabricated using a multi-material 3D printer and demonstrate potential applications. ","lang":"eng"}],"month":"07"},{"alternative_title":["ISTA Thesis"],"month":"05","acknowledged_ssus":[{"_id":"M-Shop"}],"abstract":[{"text":"Inverse design problems in fabrication-aware shape optimization are typically solved on discrete representations such as polygonal meshes. This thesis argues that there are benefits to treating these problems in the same domain as human designers, namely, the parametric one. One reason is that discretizing a parametric model usually removes the capability of making further manual changes to the design, because the human intent is captured by the shape parameters. Beyond this, knowledge about a design problem can sometimes reveal a structure that is present in a smooth representation, but is fundamentally altered by discretizing. In this case, working in the parametric domain may even simplify the optimization task. We present two lines of research that explore both of these aspects of fabrication-aware shape optimization on parametric representations.\r\n\r\nThe first project studies the design of plane elastic curves and Kirchhoff rods, which are common mathematical models for describing the deformation of thin elastic rods such as beams, ribbons, cables, and hair. Our main contribution is a characterization of all curved shapes that can be attained by bending and twisting elastic rods having a stiffness that is allowed to vary across the length. Elements like these can be manufactured using digital fabrication devices such as 3d printers and digital cutters, and have applications in free-form architecture and soft robotics.\r\n\r\nWe show that the family of curved shapes that can be produced this way admits geometric description that is concise and computationally convenient. In the case of plane curves, the geometric description is intuitive enough to allow a designer to determine whether a curved shape is physically achievable by visual inspection alone. We also present shape optimization algorithms that convert a user-defined curve in the plane or in three dimensions into the geometry of an elastic rod that will naturally deform to follow this curve when its endpoints are attached to a support structure. Implemented in an interactive software design tool, the rod geometry is generated in real time as the user edits a curve and enables fast prototyping. \r\n\r\nThe second project tackles the problem of general-purpose shape optimization on CAD models using a novel variant of the extended finite element method (XFEM). Our goal is the decoupling between the simulation mesh and the CAD model, so no geometry-dependent meshing or remeshing needs to be performed when the CAD parameters change during optimization. This is achieved by discretizing the embedding space of the CAD model, and using a new high-accuracy numerical integration method to enable XFEM on free-form elements bounded by the parametric surface patches of the model. Our simulation is differentiable from the CAD parameters to the simulation output, which enables us to use off-the-shelf gradient-based optimization procedures. The result is a method that fits seamlessly into the CAD workflow because it works on the same representation as the designer, enabling the alternation of manual editing and fabrication-aware optimization at will.","lang":"eng"}],"oa_version":"Published Version","ec_funded":1,"related_material":{"record":[{"status":"public","id":"9817","relation":"part_of_dissertation"},{"id":"7117","status":"public","relation":"part_of_dissertation"},{"relation":"dissertation_contains","status":"public","id":"13188"}]},"publication_status":"published","degree_awarded":"PhD","publication_identifier":{"issn":["2663-337X"],"isbn":["978-3-99078-031-2"]},"language":[{"iso":"eng"}],"file":[{"content_type":"application/pdf","relation":"main_file","access_level":"open_access","embargo":"2023-12-07","checksum":"cc2094e92fa27000b70eb4bfb76d6b5a","file_id":"12942","file_size":50714445,"date_updated":"2023-12-08T23:30:04Z","creator":"chafner","file_name":"thesis-hafner-2023may11-a2b.pdf","date_created":"2023-05-11T10:43:20Z"},{"file_size":265319,"date_updated":"2023-12-08T23:30:04Z","creator":"chafner","file_name":"thesis-release-form.pdf","date_created":"2023-05-11T10:43:44Z","embargo_to":"open_access","content_type":"application/pdf","relation":"source_file","access_level":"closed","file_id":"12943","checksum":"a6b51334be2b81672357b1549afab40c"}],"type":"dissertation","status":"public","_id":"12897","department":[{"_id":"GradSch"},{"_id":"BeBi"}],"file_date_updated":"2023-12-08T23:30:04Z","date_updated":"2024-01-29T10:47:51Z","supervisor":[{"last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"}],"ddc":["516","004","518","531"],"oa":1,"publisher":"Institute of Science and Technology Austria","page":"180","date_created":"2023-05-05T10:40:14Z","date_published":"2023-05-05T00:00:00Z","doi":"10.15479/at:ista:12897","year":"2023","has_accepted_license":"1","day":"05","project":[{"name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020"}],"article_processing_charge":"No","author":[{"full_name":"Hafner, Christian","last_name":"Hafner","first_name":"Christian","id":"400429CC-F248-11E8-B48F-1D18A9856A87"}],"title":"Inverse shape design with parametric representations: Kirchhoff Rods and parametric surface models","citation":{"short":"C. Hafner, Inverse Shape Design with Parametric Representations: Kirchhoff Rods and Parametric Surface Models, Institute of Science and Technology Austria, 2023.","ieee":"C. Hafner, “Inverse shape design with parametric representations: Kirchhoff Rods and parametric surface models,” Institute of Science and Technology Austria, 2023.","ama":"Hafner C. Inverse shape design with parametric representations: Kirchhoff Rods and parametric surface models. 2023. doi:10.15479/at:ista:12897","apa":"Hafner, C. (2023). Inverse shape design with parametric representations: Kirchhoff Rods and parametric surface models. Institute of Science and Technology Austria. https://doi.org/10.15479/at:ista:12897","mla":"Hafner, Christian. Inverse Shape Design with Parametric Representations: Kirchhoff Rods and Parametric Surface Models. Institute of Science and Technology Austria, 2023, doi:10.15479/at:ista:12897.","ista":"Hafner C. 2023. Inverse shape design with parametric representations: Kirchhoff Rods and parametric surface models. Institute of Science and Technology Austria.","chicago":"Hafner, Christian. “Inverse Shape Design with Parametric Representations: Kirchhoff Rods and Parametric Surface Models.” Institute of Science and Technology Austria, 2023. https://doi.org/10.15479/at:ista:12897."},"user_id":"400429CC-F248-11E8-B48F-1D18A9856A87"},{"project":[{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020"}],"article_number":"171","author":[{"full_name":"Hafner, Christian","last_name":"Hafner","first_name":"Christian","id":"400429CC-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd"}],"external_id":{"isi":["001086833300010"]},"article_processing_charge":"No","title":"The design space of Kirchhoff rods","citation":{"mla":"Hafner, Christian, and Bernd Bickel. “The Design Space of Kirchhoff Rods.” ACM Transactions on Graphics, vol. 42, no. 5, 171, Association for Computing Machinery, 2023, doi:10.1145/3606033.","apa":"Hafner, C., & Bickel, B. (2023). The design space of Kirchhoff rods. ACM Transactions on Graphics. Association for Computing Machinery. https://doi.org/10.1145/3606033","ama":"Hafner C, Bickel B. The design space of Kirchhoff rods. ACM Transactions on Graphics. 2023;42(5). doi:10.1145/3606033","short":"C. Hafner, B. Bickel, ACM Transactions on Graphics 42 (2023).","ieee":"C. Hafner and B. Bickel, “The design space of Kirchhoff rods,” ACM Transactions on Graphics, vol. 42, no. 5. Association for Computing Machinery, 2023.","chicago":"Hafner, Christian, and Bernd Bickel. “The Design Space of Kirchhoff Rods.” ACM Transactions on Graphics. Association for Computing Machinery, 2023. https://doi.org/10.1145/3606033.","ista":"Hafner C, Bickel B. 2023. The design space of Kirchhoff rods. ACM Transactions on Graphics. 42(5), 171."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","quality_controlled":"1","publisher":"Association for Computing Machinery","oa":1,"acknowledgement":"We thank the anonymous reviewers for their generous feedback, and Julian Fischer for his help in proving Proposition 1. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 715767).","doi":"10.1145/3606033","date_published":"2023-09-20T00:00:00Z","date_created":"2023-07-04T07:41:30Z","isi":1,"has_accepted_license":"1","year":"2023","day":"20","publication":"ACM Transactions on Graphics","type":"journal_article","article_type":"original","status":"public","keyword":["Computer Graphics","Computational Design","Computational Geometry","Shape Modeling"],"_id":"13188","file_date_updated":"2023-07-04T08:11:28Z","department":[{"_id":"BeBi"}],"date_updated":"2024-03-27T23:30:46Z","ddc":["516"],"month":"09","intvolume":" 42","acknowledged_ssus":[{"_id":"M-Shop"}],"abstract":[{"lang":"eng","text":"The Kirchhoff rod model describes the bending and twisting of slender elastic rods in three dimensions, and has been widely studied to enable the prediction of how a rod will deform, given its geometry and boundary conditions. In this work, we study a number of inverse problems with the goal of computing the geometry of a straight rod that will automatically deform to match a curved target shape after attaching its endpoints to a support structure. Our solution lets us finely control the static equilibrium state of a rod by varying the cross-sectional profiles along its length.\r\nWe also show that the set of physically realizable equilibrium states admits a concise geometric description in terms of linear line complexes, which leads to very efficient computational design algorithms. Implemented in an interactive software tool, they allow us to convert three-dimensional hand-drawn spline curves to elastic rods, and give feedback about the feasibility and practicality of a design in real time. We demonstrate the efficacy of our method by designing and manufacturing several physical prototypes with applications to interior design and soft robotics."}],"oa_version":"Submitted Version","related_material":{"record":[{"status":"public","id":"12897","relation":"part_of_dissertation"}]},"issue":"5","volume":42,"ec_funded":1,"publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"publication_status":"published","file":[{"file_name":"kirchhoff-rods.pdf","date_created":"2023-07-04T08:11:28Z","file_size":19635168,"date_updated":"2023-07-04T08:11:28Z","creator":"chafner","success":1,"checksum":"4954c1cfa487725bc156dcfec872478a","file_id":"13194","content_type":"application/pdf","relation":"main_file","access_level":"open_access"},{"file_name":"supp-main.pdf","date_created":"2023-07-04T07:46:28Z","title":"Supplemental Material with Proofs","file_size":420909,"date_updated":"2023-07-04T07:46:28Z","creator":"chafner","checksum":"79c9975fbc82ff71f1767331d2204cca","file_id":"13190","content_type":"application/pdf","relation":"supplementary_material","access_level":"open_access"},{"content_type":"application/pdf","access_level":"open_access","relation":"supplementary_material","checksum":"4ab647e4f03c711e1e6a5fc1eb8684db","file_id":"13191","date_updated":"2023-07-04T07:46:30Z","file_size":430086,"creator":"chafner","title":"Cheat Sheet for Notation","date_created":"2023-07-04T07:46:30Z","file_name":"supp-cheat.pdf"},{"file_id":"13192","checksum":"c0fd9a57d012046de90c185ffa904b76","access_level":"open_access","relation":"supplementary_material","content_type":"video/mp4","title":"Supplemental Video","date_created":"2023-07-04T07:46:39Z","file_name":"kirchhoff-video-final.mp4","creator":"chafner","date_updated":"2023-07-04T07:46:39Z","file_size":268088064},{"access_level":"open_access","relation":"supplementary_material","content_type":"application/x-zip-compressed","checksum":"71b00712b489ada2cd9815910ee180a9","file_id":"13193","creator":"chafner","date_updated":"2023-07-04T07:47:10Z","file_size":25790,"date_created":"2023-07-04T07:47:10Z","title":"Matlab Source Code with Example","file_name":"matlab-submission.zip"}],"language":[{"iso":"eng"}]},{"article_processing_charge":"No","author":[{"full_name":"Chen, Bin","last_name":"Chen","first_name":"Bin"},{"full_name":"Piovarci, Michael","last_name":"Piovarci","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael"},{"first_name":"Chao","last_name":"Wang","full_name":"Wang, Chao"},{"first_name":"Hans-Peter","last_name":"Seidel","full_name":"Seidel, Hans-Peter"},{"first_name":"Piotr","full_name":"Didyk, Piotr","last_name":"Didyk"},{"last_name":"Myszkowski","full_name":"Myszkowski, Karol","first_name":"Karol"},{"full_name":"Serrano, Ana","last_name":"Serrano","first_name":"Ana"}],"title":"Gloss management for consistent reproduction of real and virtual objects","citation":{"ama":"Chen B, Piovarci M, Wang C, et al. Gloss management for consistent reproduction of real and virtual objects. In: SIGGRAPH Asia 2022 Conference Papers. Vol 2022. Association for Computing Machinery; 2022. doi:10.1145/3550469.3555406","apa":"Chen, B., Piovarci, M., Wang, C., Seidel, H.-P., Didyk, P., Myszkowski, K., & Serrano, A. (2022). Gloss management for consistent reproduction of real and virtual objects. In SIGGRAPH Asia 2022 Conference Papers (Vol. 2022). Daegu, South Korea: Association for Computing Machinery. https://doi.org/10.1145/3550469.3555406","ieee":"B. Chen et al., “Gloss management for consistent reproduction of real and virtual objects,” in SIGGRAPH Asia 2022 Conference Papers, Daegu, South Korea, 2022, vol. 2022.","short":"B. Chen, M. Piovarci, C. Wang, H.-P. Seidel, P. Didyk, K. Myszkowski, A. Serrano, in:, SIGGRAPH Asia 2022 Conference Papers, Association for Computing Machinery, 2022.","mla":"Chen, Bin, et al. “Gloss Management for Consistent Reproduction of Real and Virtual Objects.” SIGGRAPH Asia 2022 Conference Papers, vol. 2022, 35, Association for Computing Machinery, 2022, doi:10.1145/3550469.3555406.","ista":"Chen B, Piovarci M, Wang C, Seidel H-P, Didyk P, Myszkowski K, Serrano A. 2022. Gloss management for consistent reproduction of real and virtual objects. SIGGRAPH Asia 2022 Conference Papers. SIGGRAPH: Computer Graphics and Interactive Techniques Conference vol. 2022, 35.","chicago":"Chen, Bin, Michael Piovarci, Chao Wang, Hans-Peter Seidel, Piotr Didyk, Karol Myszkowski, and Ana Serrano. “Gloss Management for Consistent Reproduction of Real and Virtual Objects.” In SIGGRAPH Asia 2022 Conference Papers, Vol. 2022. Association for Computing Machinery, 2022. https://doi.org/10.1145/3550469.3555406."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","project":[{"_id":"eb901961-77a9-11ec-83b8-f5c883a62027","name":"Perception-Aware Appearance Fabrication","grant_number":"M03319"}],"article_number":"35","date_created":"2023-01-12T12:03:56Z","doi":"10.1145/3550469.3555406","date_published":"2022-11-01T00:00:00Z","year":"2022","has_accepted_license":"1","publication":"SIGGRAPH Asia 2022 Conference Papers","day":"01","oa":1,"quality_controlled":"1","publisher":"Association for Computing Machinery","acknowledgement":"This work is supported by FWF Lise Meitner (Grant M 3319), European Research Council (project CHAMELEON, Grant no. 682080), Swiss National Science Foundation (Grant no. 200502), and academic gifts from Meta.","department":[{"_id":"BeBi"}],"file_date_updated":"2023-01-24T07:35:21Z","date_updated":"2023-02-13T09:15:25Z","ddc":["000"],"tmp":{"legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","short":"CC BY (4.0)"},"conference":{"start_date":"2022-12-06","end_date":"2022-12-09","location":"Daegu, South Korea","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference"},"type":"conference","status":"public","_id":"12135","volume":2022,"publication_status":"published","publication_identifier":{"isbn":["9781450394703"]},"language":[{"iso":"eng"}],"file":[{"relation":"main_file","access_level":"open_access","content_type":"application/pdf","success":1,"checksum":"f47f3215ab8bb919e3546b3438c34c21","file_id":"12351","creator":"dernst","file_size":28826826,"date_updated":"2023-01-24T07:35:21Z","file_name":"2022_ACM_SIGGRAPH_Chen.pdf","date_created":"2023-01-24T07:35:21Z"}],"scopus_import":"1","intvolume":" 2022","month":"11","abstract":[{"text":"A good match of material appearance between real-world objects and their digital on-screen representations is critical for many applications such as fabrication, design, and e-commerce. However, faithful appearance reproduction is challenging, especially for complex phenomena, such as gloss. In most cases, the view-dependent nature of gloss and the range of luminance values required for reproducing glossy materials exceeds the current capabilities of display devices. As a result, appearance reproduction poses significant problems even with accurately rendered images. This paper studies the gap between the gloss perceived from real-world objects and their digital counterparts. Based on our psychophysical experiments on a wide range of 3D printed samples and their corresponding photographs, we derive insights on the influence of geometry, illumination, and the display’s brightness and measure the change in gloss appearance due to the display limitations. Our evaluation experiments demonstrate that using the prediction to correct material parameters in a rendering system improves the match of gloss appearance between real objects and their visualization on a display device.","lang":"eng"}],"oa_version":"Published Version"},{"abstract":[{"text":"Enabling additive manufacturing to employ a wide range of novel, functional materials can be a major boost to this technology. However, making such materials printable requires painstaking trial-and-error by an expert operator,\r\nas they typically tend to exhibit peculiar rheological or hysteresis properties. Even in the case of successfully finding the process parameters, there is no guarantee of print-to-print consistency due to material differences between batches. These challenges make closed-loop feedback an attractive option where the process parameters are adjusted on-the-fly. There are several challenges for designing an efficient controller: the deposition parameters are complex and highly coupled, artifacts occur after long time horizons, simulating the deposition is computationally costly, and learning on hardware is intractable. In this work, we demonstrate the feasibility of learning a closed-loop control policy for additive manufacturing using reinforcement learning. We show that approximate, but efficient, numerical simulation is\r\nsufficient as long as it allows learning the behavioral patterns of deposition that translate to real-world experiences. In combination with reinforcement learning, our model can be used to discover control policies that outperform\r\nbaseline controllers. Furthermore, the recovered policies have a minimal sim-to-real gap. We showcase this by applying our control policy in-vivo on a single-layer, direct ink writing printer. ","lang":"eng"}],"oa_version":"Submitted Version","intvolume":" 41","month":"06","publication_status":"published","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"language":[{"iso":"eng"}],"file":[{"checksum":"27f6fe41c6ff84d50445cc9b0176d45b","file_id":"11467","success":1,"content_type":"application/pdf","access_level":"open_access","relation":"main_file","date_created":"2022-06-28T08:32:58Z","file_name":"2022_ACM_acceptedversion_Piovarci.pdf","date_updated":"2022-06-28T08:32:58Z","file_size":33994829,"creator":"dernst"}],"ec_funded":1,"volume":41,"related_material":{"link":[{"relation":"press_release","url":"https://ista.ac.at/en/news/machine-learning-3d-printing-fluids/","description":"News on ISTA website"}]},"issue":"4","_id":"11442","tmp":{"legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","short":"CC BY (4.0)"},"article_type":"original","type":"journal_article","status":"public","date_updated":"2023-05-31T12:38:21Z","ddc":["000"],"department":[{"_id":"BeBi"}],"file_date_updated":"2022-06-28T08:32:58Z","acknowledgement":"This work is graciously supported by the following grant agencies: FWF Lise Meitner (Grant M 3319), SNSF (Grant 200502), ERC Starting Grant (MATERIALIZABLE-715767), NSF (Grant IIS-181507).\r\n","oa":1,"publisher":"Association for Computing Machinery","quality_controlled":"1","year":"2022","has_accepted_license":"1","publication":"ACM Transactions on Graphics","day":"01","date_created":"2022-06-10T06:41:47Z","date_published":"2022-06-01T00:00:00Z","doi":"10.1145/3528223.3530144","article_number":"112","project":[{"name":"Perception-Aware Appearance Fabrication","grant_number":"M03319","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"},{"call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425","grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling"}],"citation":{"chicago":"Piovarci, Michael, Michael Foshey, Jie Xu, Timothy Erps, Vahid Babaei, Piotr Didyk, Szymon Rusinkiewicz, Wojciech Matusik, and Bernd Bickel. “Closed-Loop Control of Direct Ink Writing via Reinforcement Learning.” ACM Transactions on Graphics. Association for Computing Machinery, 2022. https://doi.org/10.1145/3528223.3530144.","ista":"Piovarci M, Foshey M, Xu J, Erps T, Babaei V, Didyk P, Rusinkiewicz S, Matusik W, Bickel B. 2022. Closed-loop control of direct ink writing via reinforcement learning. ACM Transactions on Graphics. 41(4), 112.","mla":"Piovarci, Michael, et al. “Closed-Loop Control of Direct Ink Writing via Reinforcement Learning.” ACM Transactions on Graphics, vol. 41, no. 4, 112, Association for Computing Machinery, 2022, doi:10.1145/3528223.3530144.","ieee":"M. Piovarci et al., “Closed-loop control of direct ink writing via reinforcement learning,” ACM Transactions on Graphics, vol. 41, no. 4. Association for Computing Machinery, 2022.","short":"M. Piovarci, M. Foshey, J. Xu, T. Erps, V. Babaei, P. Didyk, S. Rusinkiewicz, W. Matusik, B. Bickel, ACM Transactions on Graphics 41 (2022).","apa":"Piovarci, M., Foshey, M., Xu, J., Erps, T., Babaei, V., Didyk, P., … Bickel, B. (2022). Closed-loop control of direct ink writing via reinforcement learning. ACM Transactions on Graphics. Association for Computing Machinery. https://doi.org/10.1145/3528223.3530144","ama":"Piovarci M, Foshey M, Xu J, et al. Closed-loop control of direct ink writing via reinforcement learning. ACM Transactions on Graphics. 2022;41(4). doi:10.1145/3528223.3530144"},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","external_id":{"arxiv":["2201.11819"]},"article_processing_charge":"No","author":[{"id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael","full_name":"Piovarci, Michael","last_name":"Piovarci"},{"last_name":"Foshey","full_name":"Foshey, Michael","first_name":"Michael"},{"last_name":"Xu","full_name":"Xu, Jie","first_name":"Jie"},{"full_name":"Erps, Timothy","last_name":"Erps","first_name":"Timothy"},{"first_name":"Vahid","last_name":"Babaei","full_name":"Babaei, Vahid"},{"first_name":"Piotr","full_name":"Didyk, Piotr","last_name":"Didyk"},{"first_name":"Szymon","last_name":"Rusinkiewicz","full_name":"Rusinkiewicz, Szymon"},{"full_name":"Matusik, Wojciech","last_name":"Matusik","first_name":"Wojciech"},{"first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd"}],"title":"Closed-loop control of direct ink writing via reinforcement learning"},{"oa":1,"publisher":"Wiley","quality_controlled":"1","acknowledgement":"This work was supported by the Research Grants Council of the Hong Kong Special Administrative Region, China [Project No.: CUHK 14201921] and the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 715767 – MATERIALIZABLE). We thank the anonymous reviewers for their insightful feedback; Christian Hafner for proofreading and discussions; Ziqi Wang,\r\nHaisen Zhao, and Martin Hafskjold Thoresen for the helpful discussions; and the Miba Machine Shop at IST Austria for 3D printing the BUNNY and BOOMERANG models.","date_created":"2022-03-27T17:34:17Z","date_published":"2022-05-01T00:00:00Z","doi":"10.1111/cgf.14490","page":"507-519","publication":"Computer Graphics Forum","day":"01","year":"2022","isi":1,"has_accepted_license":"1","project":[{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767"}],"title":"Worst-case rigidity analysis and optimization for assemblies with mechanical joints","external_id":{"isi":["000802723900039"]},"article_processing_charge":"No","author":[{"first_name":"Zhenyuan","id":"70f0d7cf-ae65-11ec-a14f-89dfc5505b19","full_name":"Liu, Zhenyuan","orcid":"0000-0001-9200-5690","last_name":"Liu"},{"full_name":"Hu, Jingyu","last_name":"Hu","first_name":"Jingyu"},{"full_name":"Xu, Hao","last_name":"Xu","first_name":"Hao"},{"first_name":"Peng","last_name":"Song","full_name":"Song, Peng"},{"full_name":"Zhang, Ran","last_name":"Zhang","first_name":"Ran"},{"last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd"},{"last_name":"Fu","full_name":"Fu, Chi-Wing","first_name":"Chi-Wing"}],"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","citation":{"short":"Z. Liu, J. Hu, H. Xu, P. Song, R. Zhang, B. Bickel, C.-W. Fu, Computer Graphics Forum 41 (2022) 507–519.","ieee":"Z. Liu et al., “Worst-case rigidity analysis and optimization for assemblies with mechanical joints,” Computer Graphics Forum, vol. 41, no. 2. Wiley, pp. 507–519, 2022.","ama":"Liu Z, Hu J, Xu H, et al. Worst-case rigidity analysis and optimization for assemblies with mechanical joints. Computer Graphics Forum. 2022;41(2):507-519. doi:10.1111/cgf.14490","apa":"Liu, Z., Hu, J., Xu, H., Song, P., Zhang, R., Bickel, B., & Fu, C.-W. (2022). Worst-case rigidity analysis and optimization for assemblies with mechanical joints. Computer Graphics Forum. Wiley. https://doi.org/10.1111/cgf.14490","mla":"Liu, Zhenyuan, et al. “Worst-Case Rigidity Analysis and Optimization for Assemblies with Mechanical Joints.” Computer Graphics Forum, vol. 41, no. 2, Wiley, 2022, pp. 507–19, doi:10.1111/cgf.14490.","ista":"Liu Z, Hu J, Xu H, Song P, Zhang R, Bickel B, Fu C-W. 2022. Worst-case rigidity analysis and optimization for assemblies with mechanical joints. Computer Graphics Forum. 41(2), 507–519.","chicago":"Liu, Zhenyuan, Jingyu Hu, Hao Xu, Peng Song, Ran Zhang, Bernd Bickel, and Chi-Wing Fu. “Worst-Case Rigidity Analysis and Optimization for Assemblies with Mechanical Joints.” Computer Graphics Forum. Wiley, 2022. https://doi.org/10.1111/cgf.14490."},"intvolume":" 41","month":"05","scopus_import":"1","oa_version":"Submitted Version","abstract":[{"text":"We study structural rigidity for assemblies with mechanical joints. Existing methods identify whether an assembly is structurally rigid by assuming parts are perfectly rigid. Yet, an assembly identified as rigid may not be that “rigid” in practice, and existing methods cannot quantify how rigid an assembly is. We address this limitation by developing a new measure, worst-case rigidity, to quantify the rigidity of an assembly as the largest possible deformation that the assembly undergoes for arbitrary external loads of fixed magnitude. Computing worst-case rigidity is non-trivial due to non-rigid parts and different joint types. We thus formulate a new computational approach by encoding parts and their connections into a stiffness matrix, in which parts are modeled as deformable objects and joints as soft constraints. Based on this, we formulate worst-case rigidity analysis as an optimization that seeks the worst-case deformation of an assembly for arbitrary external loads, and solve the optimization problem via an eigenanalysis. Furthermore, we present methods to optimize the geometry and topology of various assemblies to enhance their rigidity, as guided by our rigidity measure. In the end, we validate our method on a variety of assembly structures with physical experiments and demonstrate its effectiveness by designing and fabricating several structurally rigid assemblies.","lang":"eng"}],"acknowledged_ssus":[{"_id":"M-Shop"}],"ec_funded":1,"volume":41,"issue":"2","language":[{"iso":"eng"}],"file":[{"access_level":"open_access","relation":"main_file","content_type":"application/pdf","checksum":"b62188b07f5c000f1638c782ec92da41","file_id":"10923","creator":"bbickel","date_updated":"2022-03-27T17:34:11Z","file_size":19601689,"date_created":"2022-03-27T17:34:11Z","file_name":"paper.pdf"}],"publication_status":"published","publication_identifier":{"issn":["0167-7055"],"eissn":["1467-8659"]},"status":"public","article_type":"original","type":"journal_article","_id":"10922","file_date_updated":"2022-03-27T17:34:11Z","department":[{"_id":"BeBi"}],"ddc":["000"],"date_updated":"2023-08-03T06:17:13Z"},{"article_type":"original","type":"journal_article","status":"public","_id":"11735","department":[{"_id":"BeBi"}],"file_date_updated":"2022-08-28T07:56:19Z","date_updated":"2023-08-03T13:21:22Z","ddc":["000"],"scopus_import":"1","month":"07","intvolume":" 41","abstract":[{"lang":"eng","text":"Interlocking puzzles are intriguing geometric games where the puzzle pieces are held together based on their geometric arrangement, preventing the puzzle from falling apart. High-level-of-difficulty, or simply high-level, interlocking puzzles are a subclass of interlocking puzzles that require multiple moves to take out the first subassembly from the puzzle. Solving a high-level interlocking puzzle is a challenging task since one has to explore many different configurations of the puzzle pieces until reaching a configuration where the first subassembly can be taken out. Designing a high-level interlocking puzzle with a user-specified level of difficulty is even harder since the puzzle pieces have to be interlocking in all the configurations before the first subassembly is taken out.\r\n\r\nIn this paper, we present a computational approach to design high-level interlocking puzzles. The core idea is to represent all possible configurations of an interlocking puzzle as well as transitions among these configurations using a rooted, undirected graph called a disassembly graph and leverage this graph to find a disassembly plan that requires a minimal number of moves to take out the first subassembly from the puzzle. At the design stage, our algorithm iteratively constructs the geometry of each puzzle piece to expand the disassembly graph incrementally, aiming to achieve a user-specified level of difficulty. We show that our approach allows efficient generation of high-level interlocking puzzles of various shape complexities, including new solutions not attainable by state-of-the-art approaches."}],"oa_version":"Submitted Version","issue":"4","volume":41,"related_material":{"link":[{"relation":"press_release","url":"https://ista.ac.at/en/news/unlocking-interlocking-riddles/","description":"News on ISTA website"}]},"ec_funded":1,"publication_identifier":{"issn":["0730-0301"],"eissn":["1557-7368"]},"publication_status":"published","file":[{"file_size":16896871,"date_updated":"2022-08-28T07:56:19Z","creator":"bbickel","file_name":"Chen-2022-High-LevelPuzzle_authorVersion.pdf","date_created":"2022-08-28T07:56:19Z","content_type":"application/pdf","relation":"main_file","access_level":"open_access","success":1,"checksum":"0b51651be45b1b33f2072bd5d2686c69","file_id":"11992"}],"language":[{"iso":"eng"}],"project":[{"call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425","grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling"}],"article_number":"150","author":[{"full_name":"Chen, Rulin","last_name":"Chen","first_name":"Rulin"},{"full_name":"Wang, Ziqi","last_name":"Wang","first_name":"Ziqi"},{"first_name":"Peng","last_name":"Song","full_name":"Song, Peng"},{"full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel","first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"}],"external_id":{"isi":["000830989200018"]},"article_processing_charge":"No","title":"Computational design of high-level interlocking puzzles","citation":{"mla":"Chen, Rulin, et al. “Computational Design of High-Level Interlocking Puzzles.” ACM Transactions on Graphics, vol. 41, no. 4, 150, Association for Computing Machinery, 2022, doi:10.1145/3528223.3530071.","ama":"Chen R, Wang Z, Song P, Bickel B. Computational design of high-level interlocking puzzles. ACM Transactions on Graphics. 2022;41(4). doi:10.1145/3528223.3530071","apa":"Chen, R., Wang, Z., Song, P., & Bickel, B. (2022). Computational design of high-level interlocking puzzles. ACM Transactions on Graphics. Association for Computing Machinery. https://doi.org/10.1145/3528223.3530071","short":"R. Chen, Z. Wang, P. Song, B. Bickel, ACM Transactions on Graphics 41 (2022).","ieee":"R. Chen, Z. Wang, P. Song, and B. Bickel, “Computational design of high-level interlocking puzzles,” ACM Transactions on Graphics, vol. 41, no. 4. Association for Computing Machinery, 2022.","chicago":"Chen, Rulin, Ziqi Wang, Peng Song, and Bernd Bickel. “Computational Design of High-Level Interlocking Puzzles.” ACM Transactions on Graphics. Association for Computing Machinery, 2022. https://doi.org/10.1145/3528223.3530071.","ista":"Chen R, Wang Z, Song P, Bickel B. 2022. Computational design of high-level interlocking puzzles. ACM Transactions on Graphics. 41(4), 150."},"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","quality_controlled":"1","publisher":"Association for Computing Machinery","oa":1,"acknowledgement":"We thank the reviewers for the valuable comments, David Gontier for sharing the source code of the baseline design approach, Christian Hafner for proofreading the paper, Keenan Crane for the 3D model of Cow, and Thingiverse for the 3D models of Moai and Owl. This work was supported by the SUTD Start-up Research Grant (Number: SRG ISTD 2019 148), the Swiss National Science Foundation (NCCR Digital Fabrication Agreement #51NF40-141853), and\r\nthe European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant Agreement No 715767 – MATERIALIZABLE).","date_published":"2022-07-22T00:00:00Z","doi":"10.1145/3528223.3530071","date_created":"2022-08-07T22:01:57Z","has_accepted_license":"1","isi":1,"year":"2022","day":"22","publication":"ACM Transactions on Graphics"},{"quality_controlled":"1","publisher":"Wiley","oa":1,"page":"435-452","doi":"10.1111/cgf.14581","date_published":"2022-09-01T00:00:00Z","date_created":"2022-08-28T18:17:01Z","isi":1,"has_accepted_license":"1","year":"2022","day":"01","publication":"Computer Graphics Forum","author":[{"first_name":"Thomas","last_name":"Alderighi","full_name":"Alderighi, Thomas"},{"last_name":"Malomo","full_name":"Malomo, Luigi","first_name":"Luigi"},{"full_name":"Auzinger, Thomas","orcid":"0000-0002-1546-3265","last_name":"Auzinger","id":"4718F954-F248-11E8-B48F-1D18A9856A87","first_name":"Thomas"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd","last_name":"Bickel","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd"},{"first_name":"Paulo","full_name":"Cignoni, Paulo","last_name":"Cignoni"},{"last_name":"Pietroni","full_name":"Pietroni, Nico","first_name":"Nico"}],"article_processing_charge":"No","external_id":{"isi":["000842638900001"]},"title":"State of the art in computational mould design","citation":{"mla":"Alderighi, Thomas, et al. “State of the Art in Computational Mould Design.” Computer Graphics Forum, vol. 41, no. 6, Wiley, 2022, pp. 435–52, doi:10.1111/cgf.14581.","apa":"Alderighi, T., Malomo, L., Auzinger, T., Bickel, B., Cignoni, P., & Pietroni, N. (2022). State of the art in computational mould design. Computer Graphics Forum. Wiley. https://doi.org/10.1111/cgf.14581","ama":"Alderighi T, Malomo L, Auzinger T, Bickel B, Cignoni P, Pietroni N. State of the art in computational mould design. Computer Graphics Forum. 2022;41(6):435-452. doi:10.1111/cgf.14581","ieee":"T. Alderighi, L. Malomo, T. Auzinger, B. Bickel, P. Cignoni, and N. Pietroni, “State of the art in computational mould design,” Computer Graphics Forum, vol. 41, no. 6. Wiley, pp. 435–452, 2022.","short":"T. Alderighi, L. Malomo, T. Auzinger, B. Bickel, P. Cignoni, N. Pietroni, Computer Graphics Forum 41 (2022) 435–452.","chicago":"Alderighi, Thomas, Luigi Malomo, Thomas Auzinger, Bernd Bickel, Paulo Cignoni, and Nico Pietroni. “State of the Art in Computational Mould Design.” Computer Graphics Forum. Wiley, 2022. https://doi.org/10.1111/cgf.14581.","ista":"Alderighi T, Malomo L, Auzinger T, Bickel B, Cignoni P, Pietroni N. 2022. State of the art in computational mould design. Computer Graphics Forum. 41(6), 435–452."},"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","scopus_import":"1","month":"09","intvolume":" 41","abstract":[{"lang":"eng","text":"Moulding refers to a set of manufacturing techniques in which a mould, usually a cavity or a solid frame, is used to shape a liquid or pliable material into an object of the desired shape. The popularity of moulding comes from its effectiveness, scalability and versatility in terms of employed materials. Its relevance as a fabrication process is demonstrated by the extensive literature covering different aspects related to mould design, from material flow simulation to the automation of mould geometry design. In this state-of-the-art report, we provide an extensive review of the automatic methods for the design of moulds, focusing on contributions from a geometric perspective. We classify existing mould design methods based on their computational approach and the nature of their target moulding process. We summarize the relationships between computational approaches and moulding techniques, highlighting their strengths and limitations. Finally, we discuss potential future research directions."}],"oa_version":"Submitted Version","issue":"6","volume":41,"publication_identifier":{"issn":["0167-7055"],"eissn":["1467-8659"]},"publication_status":"published","file":[{"content_type":"application/pdf","description":"This is the pre-peer reviewed version of the following article: Alderighi, T., Malomo, L., Auzinger, T., Bickel, B., Cignoni, P. and Pietroni, N. (2022), State of the Art in Computational Mould Design. Computer Graphics Forum, which has been published in final form at https://doi.org/10.1111/cgf.14581. This article may be used for non-commercial purposes in accordance with Wiley Terms and Conditions for Use of Self-Archived Versions.","access_level":"open_access","relation":"main_file","checksum":"c40cc8ceb7b7f0512172b883d712198e","file_id":"11994","date_updated":"2022-08-28T18:18:08Z","file_size":32480850,"creator":"bbickel","title":"pre-peer reviewed version","date_created":"2022-08-28T18:18:08Z","file_name":"star_molding_preprint.pdf"}],"language":[{"iso":"eng"}],"article_type":"original","type":"journal_article","status":"public","keyword":["Computer Graphics and Computer-Aided Design"],"_id":"11993","file_date_updated":"2022-08-28T18:18:08Z","department":[{"_id":"BeBi"}],"date_updated":"2023-08-03T13:21:55Z","ddc":["000"]},{"article_number":"708","citation":{"ieee":"P. Rao et al., “VoRF: Volumetric Relightable Faces,” in 33rd British Machine Vision Conference, London, United Kingdom, 2022.","short":"P. Rao, M. B R, G. Fox, T. Weyrich, B. Bickel, H.-P. Seidel, H. Pfister, W. Matusik, A. Tewari, C. Theobalt, M. Elgharib, in:, 33rd British Machine Vision Conference, British Machine Vision Association and Society for Pattern Recognition, 2022.","ama":"Rao P, B R M, Fox G, et al. VoRF: Volumetric Relightable Faces. In: 33rd British Machine Vision Conference. British Machine Vision Association and Society for Pattern Recognition; 2022.","apa":"Rao, P., B R, M., Fox, G., Weyrich, T., Bickel, B., Seidel, H.-P., … Elgharib, M. (2022). VoRF: Volumetric Relightable Faces. In 33rd British Machine Vision Conference. London, United Kingdom: British Machine Vision Association and Society for Pattern Recognition.","mla":"Rao, Pramod, et al. “VoRF: Volumetric Relightable Faces.” 33rd British Machine Vision Conference, 708, British Machine Vision Association and Society for Pattern Recognition, 2022.","ista":"Rao P, B R M, Fox G, Weyrich T, Bickel B, Seidel H-P, Pfister H, Matusik W, Tewari A, Theobalt C, Elgharib M. 2022. VoRF: Volumetric Relightable Faces. 33rd British Machine Vision Conference. BMVC: British Machine Vision Conference, 708.","chicago":"Rao, Pramod, Mallikarjun B R, Gereon Fox, Tim Weyrich, Bernd Bickel, Hans-Peter Seidel, Hanspeter Pfister, et al. “VoRF: Volumetric Relightable Faces.” In 33rd British Machine Vision Conference. British Machine Vision Association and Society for Pattern Recognition, 2022."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","author":[{"last_name":"Rao","full_name":"Rao, Pramod","first_name":"Pramod"},{"first_name":"Mallikarjun","full_name":"B R, Mallikarjun","last_name":"B R"},{"first_name":"Gereon","last_name":"Fox","full_name":"Fox, Gereon"},{"last_name":"Weyrich","full_name":"Weyrich, Tim","first_name":"Tim"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd","last_name":"Bickel","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385"},{"first_name":"Hans-Peter","last_name":"Seidel","full_name":"Seidel, Hans-Peter"},{"last_name":"Pfister","full_name":"Pfister, Hanspeter","first_name":"Hanspeter"},{"first_name":"Wojciech","last_name":"Matusik","full_name":"Matusik, Wojciech"},{"last_name":"Tewari","full_name":"Tewari, Ayush","first_name":"Ayush"},{"full_name":"Theobalt, Christian","last_name":"Theobalt","first_name":"Christian"},{"first_name":"Mohamed","full_name":"Elgharib, Mohamed","last_name":"Elgharib"}],"title":"VoRF: Volumetric Relightable Faces","acknowledgement":"This work was supported by the ERC Consolidator Grant 4DReply (770784).","oa":1,"publisher":"British Machine Vision Association and Society for Pattern Recognition","quality_controlled":"1","year":"2022","has_accepted_license":"1","publication":"33rd British Machine Vision Conference","day":"01","date_created":"2023-01-30T10:47:06Z","date_published":"2022-12-01T00:00:00Z","_id":"12452","conference":{"name":"BMVC: British Machine Vision Conference","start_date":"2022-11-21","end_date":"2022-11-24","location":"London, United Kingdom"},"type":"conference","status":"public","date_updated":"2023-10-31T08:40:55Z","ddc":["000"],"department":[{"_id":"BeBi"}],"file_date_updated":"2023-01-30T10:48:37Z","abstract":[{"text":"Portrait viewpoint and illumination editing is an important problem with several applications in VR/AR, movies, and photography. Comprehensive knowledge of geometry and illumination is critical for obtaining photorealistic results. Current methods are unable to explicitly model in 3D while handing both viewpoint and illumination editing from a single image. In this paper, we propose VoRF, a novel approach that can take even a single portrait image as input and relight human heads under novel illuminations that can be viewed from arbitrary viewpoints. VoRF represents a human head as a continuous volumetric field and learns a prior model of human heads using a coordinate-based MLP with separate latent spaces for identity and illumination. The prior model is learnt in an auto-decoder manner over a diverse class of head shapes and appearances, allowing VoRF to generalize to novel test identities from a single input image. Additionally, VoRF has a reflectance MLP that uses the intermediate features of the prior model for rendering One-Light-at-A-Time (OLAT) images under novel views. We synthesize novel illuminations by combining these OLAT images with target environment maps. Qualitative and quantitative evaluations demonstrate the effectiveness of VoRF for relighting and novel view synthesis even when applied to unseen subjects under uncontrolled illuminations.","lang":"eng"}],"oa_version":"Published Version","main_file_link":[{"url":"https://bmvc2022.mpi-inf.mpg.de/708/","open_access":"1"}],"scopus_import":"1","month":"12","publication_status":"published","language":[{"iso":"eng"}],"file":[{"file_name":"vorf_main.pdf","date_created":"2023-01-30T10:48:18Z","title":"VoRF: Volumetric Relightable Faces","file_size":5202710,"date_updated":"2023-01-30T10:48:18Z","creator":"bbickel","checksum":"b60b70bb48700aee709c85a69231821d","file_id":"12453","content_type":"application/pdf","relation":"main_file","access_level":"open_access"},{"date_updated":"2023-01-30T10:48:29Z","file_size":37953188,"creator":"bbickel","title":"VoRF: Volumetric Relightable Faces – SUPPLEMENTAL MATERIAL –","date_created":"2023-01-30T10:48:29Z","file_name":"vorf_supp.pdf","content_type":"application/pdf","access_level":"open_access","relation":"supplementary_material","file_id":"12454","checksum":"ce5f4ce66eaaa1590ee5df989fca6f61"},{"content_type":"video/mp4","access_level":"open_access","relation":"supplementary_material","checksum":"08aecca434b08fee75ee1efe87943718","file_id":"12455","date_updated":"2023-01-30T10:48:37Z","file_size":57855492,"creator":"bbickel","date_created":"2023-01-30T10:48:37Z","file_name":"video.mp4"}]},{"author":[{"id":"39BDC62C-F248-11E8-B48F-1D18A9856A87","first_name":"Philipp","last_name":"Velicky","full_name":"Velicky, Philipp","orcid":"0000-0002-2340-7431"},{"id":"3FB91342-F248-11E8-B48F-1D18A9856A87","first_name":"Eder","full_name":"Miguel Villalba, Eder","orcid":"0000-0001-5665-0430","last_name":"Miguel Villalba"},{"last_name":"Michalska","orcid":"0000-0003-3862-1235","full_name":"Michalska, Julia M","first_name":"Julia M","id":"443DB6DE-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Donglai","last_name":"Wei","full_name":"Wei, Donglai"},{"full_name":"Lin, Zudi","last_name":"Lin","first_name":"Zudi"},{"last_name":"Watson","full_name":"Watson, Jake","orcid":"0000-0002-8698-3823","id":"63836096-4690-11EA-BD4E-32803DDC885E","first_name":"Jake"},{"first_name":"Jakob","full_name":"Troidl, Jakob","last_name":"Troidl"},{"first_name":"Johanna","full_name":"Beyer, Johanna","last_name":"Beyer"},{"full_name":"Ben Simon, Yoav","last_name":"Ben Simon","first_name":"Yoav","id":"43DF3136-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Christoph M","id":"4DF26D8C-F248-11E8-B48F-1D18A9856A87","last_name":"Sommer","orcid":"0000-0003-1216-9105","full_name":"Sommer, Christoph M"},{"id":"425C1CE8-F248-11E8-B48F-1D18A9856A87","first_name":"Wiebke","full_name":"Jahr, Wiebke","last_name":"Jahr"},{"last_name":"Cenameri","full_name":"Cenameri, Alban","id":"9ac8f577-2357-11eb-997a-e566c5550886","first_name":"Alban"},{"first_name":"Johannes","last_name":"Broichhagen","full_name":"Broichhagen, Johannes"},{"first_name":"Seth G. N.","full_name":"Grant, Seth G. N.","last_name":"Grant"},{"last_name":"Jonas","full_name":"Jonas, Peter M","orcid":"0000-0001-5001-4804","id":"353C1B58-F248-11E8-B48F-1D18A9856A87","first_name":"Peter M"},{"id":"3E57A680-F248-11E8-B48F-1D18A9856A87","first_name":"Gaia","full_name":"Novarino, Gaia","orcid":"0000-0002-7673-7178","last_name":"Novarino"},{"full_name":"Pfister, Hanspeter","last_name":"Pfister","first_name":"Hanspeter"},{"first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd"},{"id":"42EFD3B6-F248-11E8-B48F-1D18A9856A87","first_name":"Johann G","last_name":"Danzl","full_name":"Danzl, Johann G","orcid":"0000-0001-8559-3973"}],"article_processing_charge":"No","title":"Saturated reconstruction of living brain tissue","department":[{"_id":"PeJo"},{"_id":"GaNo"},{"_id":"BeBi"},{"_id":"JoDa"}],"date_updated":"2024-03-27T23:30:20Z","citation":{"ista":"Velicky P, Miguel Villalba E, Michalska JM, Wei D, Lin Z, Watson J, Troidl J, Beyer J, Ben Simon Y, Sommer CM, Jahr W, Cenameri A, Broichhagen J, Grant SGN, Jonas PM, Novarino G, Pfister H, Bickel B, Danzl JG. Saturated reconstruction of living brain tissue. bioRxiv, 10.1101/2022.03.16.484431.","chicago":"Velicky, Philipp, Eder Miguel Villalba, Julia M Michalska, Donglai Wei, Zudi Lin, Jake Watson, Jakob Troidl, et al. “Saturated Reconstruction of Living Brain Tissue.” BioRxiv. Cold Spring Harbor Laboratory, n.d. https://doi.org/10.1101/2022.03.16.484431.","ama":"Velicky P, Miguel Villalba E, Michalska JM, et al. Saturated reconstruction of living brain tissue. bioRxiv. doi:10.1101/2022.03.16.484431","apa":"Velicky, P., Miguel Villalba, E., Michalska, J. M., Wei, D., Lin, Z., Watson, J., … Danzl, J. G. (n.d.). Saturated reconstruction of living brain tissue. bioRxiv. Cold Spring Harbor Laboratory. https://doi.org/10.1101/2022.03.16.484431","ieee":"P. Velicky et al., “Saturated reconstruction of living brain tissue,” bioRxiv. Cold Spring Harbor Laboratory.","short":"P. Velicky, E. Miguel Villalba, J.M. Michalska, D. Wei, Z. Lin, J. Watson, J. Troidl, J. Beyer, Y. Ben Simon, C.M. Sommer, W. Jahr, A. Cenameri, J. Broichhagen, S.G.N. Grant, P.M. Jonas, G. Novarino, H. Pfister, B. Bickel, J.G. Danzl, BioRxiv (n.d.).","mla":"Velicky, Philipp, et al. “Saturated Reconstruction of Living Brain Tissue.” BioRxiv, Cold Spring Harbor Laboratory, doi:10.1101/2022.03.16.484431."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","type":"preprint","status":"public","_id":"11943","related_material":{"record":[{"status":"public","id":"12470","relation":"dissertation_contains"}]},"doi":"10.1101/2022.03.16.484431","date_published":"2022-05-09T00:00:00Z","date_created":"2022-08-23T11:07:59Z","publication_status":"submitted","year":"2022","day":"09","publication":"bioRxiv","language":[{"iso":"eng"}],"publisher":"Cold Spring Harbor Laboratory","oa":1,"main_file_link":[{"open_access":"1","url":"https://doi.org/10.1101/2022.03.16.484431"}],"month":"05","abstract":[{"lang":"eng","text":"Complex wiring between neurons underlies the information-processing network enabling all brain functions, including cognition and memory. For understanding how the network is structured, processes information, and changes over time, comprehensive visualization of the architecture of living brain tissue with its cellular and molecular components would open up major opportunities. However, electron microscopy (EM) provides nanometre-scale resolution required for full in-silico reconstruction1–5, yet is limited to fixed specimens and static representations. Light microscopy allows live observation, with super-resolution approaches6–12 facilitating nanoscale visualization, but comprehensive 3D-reconstruction of living brain tissue has been hindered by tissue photo-burden, photobleaching, insufficient 3D-resolution, and inadequate signal-to-noise ratio (SNR). Here we demonstrate saturated reconstruction of living brain tissue. We developed an integrated imaging and analysis technology, adapting stimulated emission depletion (STED) microscopy6,13 in extracellularly labelled tissue14 for high SNR and near-isotropic resolution. Centrally, a two-stage deep-learning approach leveraged previously obtained information on sample structure to drastically reduce photo-burden and enable automated volumetric reconstruction down to single synapse level. Live reconstruction provides unbiased analysis of tissue architecture across time in relation to functional activity and targeted activation, and contextual understanding of molecular labelling. This adoptable technology will facilitate novel insights into the dynamic functional architecture of living brain tissue."}],"oa_version":"Preprint"}]