

{"id":173,"date":"2021-05-29T20:06:55","date_gmt":"2021-05-29T18:06:55","guid":{"rendered":"https:\/\/project.inria.fr\/fungraph\/?page_id=173"},"modified":"2023-10-12T22:17:22","modified_gmt":"2023-10-12T20:17:22","slug":"code-and-data","status":"publish","type":"page","link":"https:\/\/project.inria.fr\/fungraph\/code-and-data\/","title":{"rendered":"Code and Data"},"content":{"rendered":"<p>We have published code and data for most of our projects.<\/p>\n<p>For most Image-Based Rendering project, our code is distributed as a project of SIBR (<a href=\"http:\/\/sibr.gitlabpages.inria.fr\">http:\/\/sibr.gitlabpages.inria.fr<\/a>); datasets are usually given in the corresponding space on the FUNGRAPH fileserver.<\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2021\/PLRD21\/\">Hybrid Image-based Rendering for Free-view Synthesis<\/a>&#8221; code and data are provided on the project page: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/hybrid-ibr\/\">https:\/\/repo-sam.inria.fr\/fungraph\/hybrid-ibr\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"https:\/\/www-sop.inria.fr\/reves\/Basilic\/2020\/RLPWSD20\/\">Glossy Probe Reprojection for Interactive Global Illumination<\/a>&#8221; code and data are provided on the project page:\u00a0<a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/synthetic-probes\/\">https:\/\/repo-sam.inria.fr\/fungraph\/synthetic-probes\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2019\/PGZED19\/\">Multi-view relighting using geometry and deep learning<\/a>&#8220;, code and data are provided on the project page: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/deep-relighting\/\">https:\/\/repo-sam.inria.fr\/fungraph\/deep-relighting\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2020\/RPHD20\/\">Image-Based Rendering of Cars using Semantic Labelling<\/a>&#8220;, code and data are provided on the project page:\u00a0 <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/ibr-cars-semantic\/\">https:\/\/repo-sam.inria.fr\/fungraph\/ibr-cars-semantic\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2019\/DADDB19\/\">Flexible SVBRDF Capture with a Multi-Image Deep Network<\/a>&#8220;, code and data are provided on the project page: <a href=\"https:\/\/team.inria.fr\/graphdeco\/fr\/projects\/multi-materials\/\">https:\/\/team.inria.fr\/graphdeco\/fr\/projects\/multi-materials\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"https:\/\/www-sop.inria.fr\/reves\/Basilic\/2020\/DDB20\/\">Guided Fine-Tuning for Large-Scale Material Transfer<\/a>&#8220;, code and data are provided on the project page: <a href=\"https:\/\/team.inria.fr\/graphdeco\/fr\/projects\/large-scale-materials\/\">https:\/\/team.inria.fr\/graphdeco\/fr\/projects\/large-scale-materials\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2020\/DGJND20\/\">Practical Product Path Guiding Using Linearly Transformed Cosines<\/a>&#8221; code is available here: <a href=\"https:\/\/gitlab.inria.fr\/sdiolatz\/practical-product-path-guiding\">https:\/\/gitlab.inria.fr\/sdiolatz\/practical-product-path-guiding<\/a><\/p>\n<p>For the publication\u00a0<a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/freestylegan\/\">FreeStyleGAN: Free-view Editable Portrait Rendering with the Camera Manifold<\/a> code is available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/freestylegan\/\">https:\/\/gitlab.inria.fr\/fungraph\/freestylegan<\/a><\/p>\n<p>For the publication\u00a0 &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2022\/DPD22\/\">Active Exploration for Neural Global Illumination of Variable Scenes<\/a>&#8221; code and data are available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/active-exploration\/\">https:\/\/repo-sam.inria.fr\/fungraph\/active-exploration\/<\/a><\/p>\n<p>For the publication\u00a0 &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2022\/PRBD22\/\">Deep scene-scale material estimation from multi-view indoor captures <\/a> &#8221; code and data are available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/deep_multiview_scene_materials\/\">https:\/\/repo-sam.inria.fr\/fungraph\/deep_multiview_scene_materials\/<\/a><\/p>\n<p>For the publication\u00a0 &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2022\/KLRJD22\/\">Neural Point Catacaustics for Novel-View Synthesis of Reflections<\/a>&#8221; code and data are available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/neural_catacaustics\/\">https:\/\/repo-sam.inria.fr\/fungraph\/neural_catacaustics\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2023\/JKKDLD23\/\">NeRFshop: Interactive Editing of Neural Radiance Fields<\/a>&#8221; code and data are available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/nerfshop\/\">https:\/\/repo-sam.inria.fr\/fungraph\/nerfshop\/<\/a><\/p>\n<p>For the publication &#8220;<a href=\"http:\/\/www-sop.inria.fr\/reves\/Basilic\/2023\/KKLD23\/\">3D Gaussian Splatting for Real-Time Radiance Field Rendering<\/a>&#8221; code and data are available here: <a href=\"https:\/\/repo-sam.inria.fr\/fungraph\/3d-gaussian-splatting\/\">https:\/\/repo-sam.inria.fr\/fungraph\/3d-gaussian-splatting\/<\/a><\/p>\n\n\n<p><\/p>","protected":false},"excerpt":{"rendered":"<p>We have published code and data for most of our projects. For most Image-Based Rendering project, our code is distributed as a project of SIBR (http:\/\/sibr.gitlabpages.inria.fr); datasets are usually given in the corresponding space on the FUNGRAPH fileserver. For the publication &#8220;Hybrid Image-based Rendering for Free-view Synthesis&#8221; code and data\u2026<\/p>\n<p> <a class=\"continue-reading-link\" href=\"https:\/\/project.inria.fr\/fungraph\/code-and-data\/\"><span>Continue reading<\/span><i class=\"crycon-right-dir\"><\/i><\/a> <\/p>\n","protected":false},"author":376,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-173","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/pages\/173","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/users\/376"}],"replies":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/comments?post=173"}],"version-history":[{"count":14,"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/pages\/173\/revisions"}],"predecessor-version":[{"id":219,"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/pages\/173\/revisions\/219"}],"wp:attachment":[{"href":"https:\/\/project.inria.fr\/fungraph\/wp-json\/wp\/v2\/media?parent=173"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}