CINXE.COM

Wei-Chiu Ma, Cornell

<!doctype html> <html> <head> <meta http-equiv="Content-Type" content="text/html; charset=utf-8" /> <title>Wei-Chiu Ma, Cornell</title> <style> h1 { padding : 0; margin : 0; } body { padding : 20px 0; font-family : Arial; font-size : 16px; background-image : url("img/bg.png"); } #container { width : 900px; margin : 0 auto; background-color : #fff; padding : 50px; text-align: left; box-shadow: 0px 0px 10px #999;; } #me { margin-left : 25px; border : 0 solid black; float : right; margin-bottom : 0; margin-right:25px;} #content { display : block; margin-right : 275px;} a { text-decoration : none; } a:hover { text-decoration : underline; } a:link,a:visited {color: #1367a7;} a.invisible { color : inherit; text-decoration : inherit; } .publogo { margin-top : 20px; margin-right : 10px; float : left; border : 0; width: 200px; vertical-align: middle;} .publication { clear : left; padding-bottom : 10px; line-height:22px;} .codelogo { margin-right : 10px; float : left; border : 0;} .code { clear : left; padding-bottom : 10px; vertical-align :middle;} .code .download a { display : block; margin : 0 15px; float : left;} #simpsons { margin : 5px auto; text-align : center; color : #B7B7B7; } span.highlight, span.highlight a:link, span.highlight a:visited{color:#BB2222} span.collaborator, span.collaborator a:link, span.collaborator a:visited{color:#666666} </style> <script type="text/javascript"> function showPubs(id) { if (id == 0) { document.getElementById('pubs').innerHTML = document.getElementById('pubs_selected').innerHTML; document.getElementById('select0').style = 'text-decoration:underline;color:#000000'; document.getElementById('select1').style = ''; // document.getElementById('select2').style = ''; } else if (id == 1) { document.getElementById('pubs').innerHTML = document.getElementById('pubs_by_date').innerHTML; document.getElementById('select1').style = 'text-decoration:underline;color:#000000'; document.getElementById('select0').style = ''; // document.getElementById('select2').style = ''; } // else { // document.getElementById('pubs').innerHTML = document.getElementById('pubs_by_topic').innerHTML; // document.getElementById('select2').style = 'text-decoration:underline;color:#000000'; // document.getElementById('select0').style = ''; // document.getElementById('select1').style = ''; // } } </script> </head> <body> <!--#F2F2F2--> <center> <div id="container"> <!-- <img src="img/Wei-Chiu-Ma.jpg" align = "right" width = 200> --> <img src="img/weichiu-mit-min.PNG" align = "right" width = 200> <div id="content"> <h1>Wei-Chiu Ma 馬惟九</h1> <p> <!-- Incoming Assistant Professor (Fall 2024)<br> Department of Computer Science<br> Cornell University<br><br> --> <!-- Young Investigator/Postdoc @ AI2/UW<br> --> Assistant Professor @ Cornell CS<br> <br> <!-- Ph.D. Student<br> Department of Electrical Engineering and Computer Science<br> Massachusetts Institute of Technology<br><br> --> <!-- Incoming Assistant Professor<br> Department of Computer Science<br> Cornell University<br><br> --> <a href="mailto:wm347@cornell.edu">Email</a> / <a href="https://scholar.google.com/citations?user=SVIdh6AAAAAJ&hl=en#">Google scholar</a> / <a href="https://twitter.com/weichiuma">Twitter</a> / <a href="#misc">Misc.</a> / <a href="#pro-bono">Pro Bono</a></p> <h2>About Me</h2> <!--<p>I am a Ph.D. student at <a href="http://web.mit.edu">Massachusetts Institute of Technology (MIT)</a>. I work at <a href="http://csail.mit.edu">Computer Science and Artificial Intelligence Laboratory (CSAIL)</a> where I am advised by <a href="http://web.mit.edu/torralba/www/">Antonio Torralba</a>. My research interests mainly lie in the intersection of computer vision and machine learning, with particular interests in deep learning and high-level vision tasks such as human action understanding and geo-localization.</p> --> <p> <!-- I am a Young Investigator/Postdoc at <a href="https://allenai.org">AI2</a>/<a href="https://www.washington.edu">University of Washington</a>, as well as an incoming Assistant Professor of Computer Science at <a href="http://cornell.edu">Cornell University</a>. --> I am an Assistant Professor of Computer Science at <a href="http://cornell.edu">Cornell University</a>. <!-- I am a final year Ph.D. candidate at <a href="http://web.mit.edu">MIT</a> working with <a href="http://web.mit.edu/torralba/www/">Antonio Torralba</a> (aka <a href="http://bit.ly/the-great-torralba">the Great Torralba</a>) and <a href="http://www.cs.toronto.edu/~urtasun/">Raquel Urtasun</a>. I am also a part-time Senior Research Scientist at <a href="https://waabi.ai">Waabi</a>. My research is supported by a <a href="https://www.siebelscholars.com">Siebel Scholarship</a>. --> </p> <p> My research lies at the intersection of 3D/4D computer vision and robotics. I am interested in building AI systems that can <b>understand</b>, <b>reconstruct</b>, and <b>re-simulate</b> our dynamic world, leveraging these capabilities to enable more robust autonomous systems or advance entertainment applications. <!-- Specifically, I am interested in developing robust computational tools that allow us to <b>model</b>, <b>reconstruct</b>, and <b>simulate</b> the dynamic world from sparse, noisy, and unconstrained sensory data. --> <!-- I am interested in computer vision and robotics, especially 3D vision and self-driving vehicles. Over the past few years, I have worked on various topics, such as robot localization, static 3D reconstruction, dynamic motion estimation, large-scale 3D generation, and closed-loop sensor simulation. I examine these tasks not only in controlled settings, but also in sparse, noisy, and sometimes extreme real-world settings in which the models will be deployed. --> </p> <!--<p> I am a computer vision fanatic. I am interested in a wide array of topics, ranging from low-level vision to high-level vision, and their connections to autonomous systems. Recently I have been thinking more from the 3D perspective. </p>--> <!--<p> I also work closely with <a href="http://www.cs.toronto.edu/~urtasun/">Raquel Urtasun</a> at <a href="https://eng.uber.com/tag/uber-atg-toronto/">Uber ATG Toronto</a> on self-driving problems.</p>--> <!-- <p>I am a Ph.D. student at <a href="http://web.mit.edu">MIT</a>. I work at <a href="http://csail.mit.edu">Computer Science and Artificial Intelligence Laboratory (CSAIL)</a> where I am advised by <a href="http://web.mit.edu/torralba/www/">Antonio Torralba</a>. My research interests mainly lie in the intersection of computer vision and machine learning. </p> --> <p> Prior to joining Cornell, I was a Young Investigator/Postdoc at <a href="https://allenai.org">AI2</a>/<a href="https://www.washington.edu">University of Washington</a>. I received my Ph.D. from <a href="http://web.mit.edu">MIT</a>, where I worked with <a href="http://web.mit.edu/torralba/www/">Antonio Torralba</a> (aka <a href="http://bit.ly/the-great-torralba">the Great Torralba</a>) and <a href="http://www.cs.toronto.edu/~urtasun/">Raquel Urtasun</a>. Previously, I was a Senior Research Scientist at <a href="https://eng.uber.com/tag/uber-atg-toronto/">Uber ATG R&D</a> and <a href="https://waabi.ai">Waabi</a> working on self-driving vehicles. I completed my M.S. in Robotics at <a href="http://www.cs.cmu.edu/">Carnegie Mellon University (CMU)</a>, where I was advised by <a href="http://www.cs.cmu.edu/~kkitani/">Kris M. Kitani</a>.</p> <p><span class="highlight"><b>Prospective students:</b></span> I am always looking for motivated and talented students! If you are interested in collaborating or joining my group as a PhD/MS/Undergrad student or intern, please read <a href="prospective.html">this</a>. </p> <!-- <p><span class="highlight"><b>Update:</b></span> I will be joining the <a href="http://www.cs.cornell.edu/">Computer Science Department</a> at <a href="http://cornell.edu">Cornell University</a> as an Assistant Professor in Fall 2024. I am always looking for motivated and talented students! If you are interested in collaborating or joining my group as a PhD/MS/Undergrad student or intern, please read <a href="prospective/html">this</a>. </p> --> <!--<h3>Pro-bono office hour</h3> <p>I have decided to commit 1~2 hours every week to provide guidance, suggestions, and/or mentorships for students from underrepresented groups or whoever is in need. Please fill in this <a href="https://docs.google.com/forms/d/1uYA95BMyFjnvY2-XDbwSAlTz79Ip55cytzrNoROS9ko/edit">form</a>.</p> --> <!-- <p>I completed my M.S. in Robotics at <a href="http://www.cs.cmu.edu/">Carnegie Mellon University (CMU)</a> where I was advised by <a href="http://www.cs.cmu.edu/~kkitani/">Kris M. Kitani</a>. During my Master's study, I also visited <a href="https://www.utoronto.ca">University of Toronto (UofT)</a> and worked closely with <a href="http://www.cs.toronto.edu/~urtasun/">Raquel Urtasun</a> and <a href="http://www.cs.utoronto.ca/~fidler/">Sanja Fidler</a>. Before coming to the US, I obtained my B.S. in Electrical Engineering at <a href="http://www.ntu.edu.tw">National Taiwan University (NTU)</a>.</p> --> <br> <!--<h6>Current/Past Affiliations</h6> --> <center> <table border="0" cellpadding="0" cellspacing="0" width="900" align="center" bgcolor="#FFFFFF"> <tr> <td width="20"></td> <!-- <td width="120" align="center" valign="middle"><img src="img/logo_mit.gif" height="50" /></td> <td width="120" align="center" valign="middle"><img src="img/logo_mit.gif" height="50" /></td> --> <!-- <td width="90" align="center" valign="middle"><img src="img/logo_ai2.png" height="37.5" /></td> <td width="90" align="center" valign="middle"><img src="img/logo_uw.png" height="37.5" /></td> --> <!-- <td width="90" align="center" valign="middle"><img src="img/logo_mit.gif" height="37.5" /></td> <td width="90" align="center" valign="middle"><img src="img/waabi.gif" height="60" /></td> <td width="90" align="center" valign="middle"><img src="img/logo_uber_atg.jpg" height="52.5" /></td> <td width="90" align="center" valign="middle"><img src="img/logo_cmu_1.png" height="52.5" /></td> <td width="90" align="center" valign="middle"><img src="img/logo_uoft.jpg" height="60" /></td> <td width="90" align="center" valign="middle"><img src="img/logo_ntu.jpg" height="60" /></td> --> <td width="120" align="center" valign="middle"><img src="img/logo_mit.gif" height="50" /></td> <td width="120" align="center" valign="middle"><img src="img/waabi.gif" height="80" /></td> <td width="120" align="center" valign="middle"><img src="img/logo_uber_atg.jpg" height="70" /></td> <td width="120" align="center" valign="middle"><img src="img/logo_cmu_1.png" height="70" /></td> <td width="120" align="center" valign="middle"><img src="img/logo_uoft.jpg" height="80" /></td> <td width="120" align="center" valign="middle"><img src="img/logo_ntu.jpg" height="80" /></td> <td width="20"></td> </tr> <!-- <tr> <td width="20"></td> <td width="120" align="center" valign="middle">2016 - Present</td> <td width="120" align="center" valign="middle">2016 - Present</td> <td width="120" align="center" valign="middle">2014 - 2016</td> <td width="120" align="center" valign="middle">Summer 2015, 2016</td> <td width="120" align="center" valign="middle">2009 - 2013</td> <td width="20" align="center" valign="middle"></td> <td width="20"></td> </tr> --> <tr> <td colspan = 7> <p>&nbsp;</p> <h2>Recent News</h2> <ul> <!-- <li><sup><font color="red"><b>NEW</b></font></sup> <b>Seminar</b>: I co-organize the <a href="https://sites.google.com/view/visionseminar">MIT Vision and Graphics Seminar.</a> Please reach out if you are interested in presenting.</li> --> <li><sup><font color="red"><b>NEW</b></font></sup> <b>Pro bono</b>: I will be hosting pro bono office hours starting 2021. Please check out <a href="#pro-bono">here</a> for more details.</li> <!-- <li><sup><font color="red"><b>NEW</b></font></sup> <b>Career update</b>: I will be joining the <a href="http://www.cs.cornell.edu/">Computer Science Department</a> at <a href="http://cornell.edu">Cornell University</a> as an Assistant Professor in Fall 2024! In the meantime, I will visit <a href="https://allenai.org">AI2</a> and the <a href="https://www.washington.edu">University of Washington</a>!</li> --> <br> <li><sup><font color="red"><b>NEW</b></font></sup> <b>CVPR Workshop</b>: We will be organizing the second workshop on<a href="https://syndata4cv.github.io/" target=""> Synthetic Data for Computer Vision</a> at CVPR 2025. Stay tuned for more details!</li> <li><sup><font color="red"><b>NEW</b></font></sup> <b>CVPR Workshop</b>: We will be organizing the first workshop on<a href="" target=""> Agent in Interaction, from Humans to Robots</a> at CVPR 2025. Stay tuned for more details!</li> <!-- <li><sup><font color="red"><b>NEW</b></font></sup> <b>ECCV Workshop</b>: We are organizing the <a href="https://3d-in-the-wild.github.io/" target=""> Wild3D: 3D Modeling, Reconstruction, and Generation in the Wild</a> workshop at ECCV 2024. Submit and present your cool work <a href="https://openreview.net/group?id=thecvf.com/ECCV/2024/Workshop/Wild3D#tab-recent-activity" target="">here</a>!</li> --> <br> <li><sup><font color="red"><b>NEW</b></font></sup> <b>Dec. 2024</b>: We released the 360-1M dataset! Check them out <a href="https://github.com/MattWallingford/360-1M" target="">here</a>! <li><sup><font color="red"><b>NEW</b></font></sup> <b>Dec. 2024</b>: Can AI system generate 3D worlds from a single image? Yes! <a href="https://mattwallingford.github.io/ODIN/" target="">All you need is the RIGHT data!</a> <li><sup><font color="red"><b>NEW</b></font></sup> <b>Oct. 2024</b>: We organized the <a href="https://3d-in-the-wild.github.io/" target="">3D Modeling, Reconstruction, and Generation in the Wild</a> workshop at ECCV 2024!</li> <li><b>Jul. 2024</b>: Can Multimodal LLMs (GPT-4V) perceive the world as humans do? See our <a href="https://zeyofu.github.io/blink/" target="">ECCV paper</a> for answers!</li> <li><b>Apr. 2024</b>: How can synthetic data benefit computer vision? Attend our CVPR <a href="https://syndata4cv.github.io/" target="">SynData4CV</a> workshop to find out!</li> <li><b>Mar. 2024</b>: Check out our latest effort on building <a href="https://video2game.github.io/" target="">realistic, interactive, and game-engine compatible digital twins</a>!</li> <li><b>Sep. 2023</b>: Two papers on in-the-wild/extreme 3D inverse graphics accepted to NeurIPS 2023!</li> <li><b>Sep. 2023</b>: I am now Dr. Ma! Thank you, Antonio and Raquel, for your guidance and tremendous support! </li> <!-- I could not be more fortunate or express enough gratitude for what you have done for me!</li> --> <li><b>Apr. 2023</b>: Our work on <a href='https://virtual-correspondence.github.io'>extreme-view geometry</a> is featured on <a href='https://youtu.be/y2BVTW09vck?t=360'>Vox</a>! Check it out!</li> <li><b>Apr. 2023</b>: Selected as a <a href='https://risingstars.linklab.virginia.edu/2023/'>Cyber-Physical Systems (CPS) rising star</a>!</li> <li><b>Mar. 2023</b>: Check out our latest effort on closed-loop sensor simulation, LiDAR generation, and thermal imaging!</li> <li><b>Oct. 2022</b>: Gave a talk at CMU, Columbia, and UIUC on in-the-wild 3D modeling, generation, and simulation!</li> <li><b>Sep. 2022</b>: Check out our latest effort on 3D scene generation, sensor simulation, and neural fields for manipulation!</li> <li><b>Aug. 2022</b>: Selected as a <a href="https://www.businesswire.com/news/home/20220922005006/en/Siebel-Scholars-Foundation-Announces-Class-of-2023">Siebel Scholar</a>!</li> <li> <b>Apr. 2022</b>: Gave a talk at Harvard on exploiting high-level vision for level-vision!</li> <li> <b>Mar. 2022</b>: Our work on extreme-view 3D reconsturction and planar neural field are accepted to CVPR 2022!</li> <li> <b>Jul. 2021</b>: BARF! Train your own NeRF from a collection of images without knowing camera poses!</li> <li> <b>Mar. 2021</b>: Check out our latest effort on simulating pedestrians in the wild at CoRL and CVPR!</li> <li> <b>Jul. 2020</b>: Four papers (two spotlight) accepted to ECCV 2020! Stay tuned!</li> <li> <b>Mar. 2020</b>: Our work on LiDAR simulation and instance segmentation are accepted to CVPR 2020!</li> <li> <b>Oct. 2019</b>: The source code of our real-time stereo algorithm is available now! Make sure to check out the amzaing differentiable PatchMatch module!</li> <li> <b>Jul. 2019</b>: Three papers accepted to ICCV 2019! Details coming soon!</li> <li> <b>Jun. 2019</b>: Our paper on light-weight localization is accepted to IROS 2019!</li> <li> <b>Mar. 2019</b>: Two papers (scene flow and road boundary extraction) accepted to CVPR 2019!</li> <li> <b>Jul. 2018</b>: Our paper on Unsupervised Intrinsic Decomposition is accepted to ECCV 2018!</li> <li><b>Mar. 2018</b>: 3 papers accepted to CVPR 2018!</li> <li> <b>Apr. 2017</b>: Our book chapter on Activity Forecasting is published! Check it out!</li> <li><b>Mar. 2017</b>: Our work (a game-theoretic approach to multi-agent activity forecasting) is accepted to CVPR 2017!</li> <li><b>Jan. 2017</b>: Our work (self-localization for autonomous vehicles) is accepted to ICRA 2017!</li> <!--<li><b>Sep. 2016</b>: I joined the EECS department at MIT as a Ph.D. student!</li> <li><b>May. - Aug. 2016</b>: I visited University of Toronto again this summer!</li>--> <li><b>May. 2016</b>: I graduated from CMU --- thanks Kris for all your support in the past two years!</li> </ul> </td> </tr> <tr> <td colspan = 7> <p>&nbsp;</p> <h2 style="display:inline;">Research Projects</h2> &nbsp (<a href="" id="select0" onclick="showPubs(0); return false;">show selected</a> / <a href="" id="select1" onclick="showPubs(1); return false;">show by date</a>) <!-- / <a href="" id="select2" onclick="showPubs(2); return false;">show by topic</a>) --> <!--<div style="top:-10px; position:relative;">--> <table border="0" cellpadding="0" cellspacing="0" width="900" align="center" bgcolor="#FFFFFF" id="pubs"></table> <script id="pubs_selected" language="text"> <!-- Odin --> <tr> <td> <img border=0 src="img/odin.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>From an Image to a Scene: Learning to Imagine the World from a Million 360° Videos</strong> <br> <span class="collaborator">Matthew Wallingford, </span> <span class="collaborator">Anand Bhattad, </span> <span class="collaborator">Aditya Kusupati, </span> <span class="collaborator">Vivek Ramanujan, </span> <span class="collaborator">Matt Deitke, </span> <span class="collaborator">Sham Kakade, </span> <span class="collaborator">Aniruddha Kembhavi, </span> <span class="collaborator">Roozbeh Mottaghi, </span> Wei-Chiu Ma, <span class="collaborator">Ali Farhadi</span> <br> <a href="" target=""><b>NeurIPS 2024</b></a> / <a href="https://mattwallingford.github.io/ODIN/" target="">project page</a> / <a href="https://arxiv.org/pdf/2412.07770" target="">arXiv</a> / <a href="https://huggingface.co/datasets/mwallingford/360-1M/tree/main" target="">huggingface</a> </div> </td> </tr> <!-- eval3d --> <!-- <tr> <td> <img border=0 src="img/eval3d.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Eval3D: Interpretable and Fine-grained Evaluation for 3D Generation</strong> <br> <span class="collaborator">Shivam Duggal*, </span> <span class="collaborator">Yushi Hu*, </span> <span class="collaborator">Oscar Michel, </span> <span class="collaborator">Aniruddha Kembhavi, </span> <span class="collaborator">William T. Freeman, </span> <span class="collaborator">Noah A. Smith, </span> <span class="collaborator">Ranjay Krishna, </span> <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Ali Farhadi, </span> Wei-Chiu Ma <br> <a href="" target=""><b>arXiv coming soon!</b></a> </div> </td> </tr> --> <!-- BLINK --> <tr> <td> <img border=0 src="img/blink.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>BLINK: Multimodal Large Language Models Can See but Not Perceive</strong> <br> <span class="collaborator">Xingyu Fu*, </span> <span class="collaborator">Yushi Hu*, </span> <span class="collaborator">Bangzheng Li, </span> <span class="collaborator">Yu Feng, </span> <span class="collaborator">Haoyu Wang, </span> <span class="collaborator">Xudong Lin, </span> <span class="collaborator">Dan Roth, </span> <span class="collaborator">Noah A. Smith, </span> Wei-Chiu Ma<sup>&#8224;</sup>, <span class="collaborator">Ranjay Krishna<sup>&#8224;</sup></span> <br> <a href="" target=""><b>ECCV 2024</b></a> / <a href="https://zeyofu.github.io/blink/" target="">project page</a> / <a href="https://arxiv.org/pdf/2404.12390" target="">paper</a> / <a href="https://huggingface.co/datasets/BLINK-Benchmark/BLINK" target="">dataset</a> / <a href="https://eval.ai/web/challenges/challenge-page/2287/overview" target="">Eval AI</a> / <a href="https://github.com/zeyofu/BLINK_Benchmark" target="">code</a> </div> </td> </tr> <!-- Video2Game --> <tr> <td> <img border=0 src="img/video2game.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Video2Game: Real-time, Interactive, Realistic and Browser-Compatible Environment from a Single Video</strong> <br> <span class="collaborator">Hongchi Xia, </span> <span class="collaborator">Zhi-Hao Lin, </span> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang</span> <br> <a href="" target=""><b>CVPR 2024</b></a> / <a href="https://video2game.github.io/" target="">project page</a> / <a href="https://arxiv.org/pdf/2404.09833" target="">paper</a> / <a href="https://video2game.github.io/src/garden/index.html" target="">shooting demo</a> / <a href="https://github.com/video2game/video2game" target="">code</a> </div> </td> </tr> <!-- UniSim --> <tr> <td> <img border=0 src="img/unisim.mp4" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>UniSim: A Neural Closed-Loop Sensor Simulator</strong> <br> <span class="collaborator">Ze Yang*, </span> <span class="collaborator">Yun Chen*, </span> <span class="collaborator">Jingkang Wang*, </span> <span class="collaborator">Sivabalan Manivasagam*, </span> Wei-Chiu Ma, <span class="collaborator">Joyce Anqi Yang, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CVPR 2023</b></a> / <a href="https://waabi.ai/unisim/" target="">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2023/papers/Yang_UniSim_A_Neural_Closed-Loop_Sensor_Simulator_CVPR_2023_paper.pdf" target="">paper</a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/unisim/unisim_final_v2_4k.mp4" target="">4K demo</a> / <a href="https://waabi.ai/wp-content/uploads/2023/05/UniSim-video_compressed.mp4" target="">video (8 mins)</a> <br> <span class="highlight"><b>Highlight presentation</b></span> </div> </td> </tr> <!-- UltraLidar --> <tr> <td> <img border=0 src="img/lidar-gen.mp4" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>UltraLiDAR: Learning Compact Representations for LiDAR Completion and Generation</strong> <br> <span class="collaborator">Yuwen Xiong, </span> Wei-Chiu Ma, <span class="collaborator">Jingkang Wang, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CVPR 2023</b></a> / <a href="https://waabi.ai/ultralidar/" target="">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2023/papers/Xiong_Learning_Compact_Representations_for_LiDAR_Completion_and_Generation_CVPR_2023_paper.pdf" target="">paper</a> / <a href="https://waabi.ai/wp-content/uploads/2023/05/UltraLidar-video.mov" target="">video (1 min)</a> </div> </td> </tr> <!-- SGAM --> <tr> <td> <img border=0 src="papers/neurips22-sgam/sgam.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>SGAM: Building a Virtual 3D World through Simultaneous Generation and Mapping</strong> <br> <span class="collaborator">Yuan Shen, </span> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang</span> <br> <a href="" target=""><b>NeurIPS 2022</b></a> / <a href="https://yshen47.github.io/sgam/" target="">project page</a> / <a href="https://openreview.net/pdf?id=17KCLTbRymw" target="">paper</a> / <a href="https://github.com/yshen47/SGAM" target="">code</a> </div> </td> </tr> <!-- CADSim --> <tr> <td> <img border=0 src="papers/corl22-cad-sim/cad-sim.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>CADSim: Robust and Scalable in-the-wild 3D Reconstruction for Realistic and Controllable Sensor Simulation</strong> <br> <span class="collaborator">Jingkang Wang, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Yun Chen, </span> <span class="collaborator">Ze Yang, </span> <span class="collaborator">Ioan Andrei Bârsan, </span> <span class="collaborator">Joyce Anqi Yang, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CoRL 2022</b></a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/cadsim.html" target="">project page</a> / <a href="https://openreview.net/pdf?id=Mp3Y5jd7rnW" target="">paper</a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/cadsim/cadsim_corl_denoise.mp4" target="">video</a> </div> </td> </tr> <!-- VC --> <tr> <td> <img border=0 src="img/vc-3.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Virtual Correspondence: Humans as a Cue for Extreme-View Geometry</strong> <br> Wei-Chiu Ma, <span class="collaborator">Anqi Joyce Yang, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Raquel Urtasun, </span> <span class="collaborator">Antonio Torralba</span> <br> <a href="" target="_new"><b>CVPR 2022</b></a> / <a href="virtual-correspondence" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2206.08365.pdf" target="_new">paper</a> / <a href="https://virtual-correspondence.github.io/img/vc_demo.mp4" target="_new">video (1.5 mins)</a> / <a href="https://youtu.be/W9odd2F2Bx4" target="_new">video (5 mins)</a> / <a href="https://news.mit.edu/2022/seeing-whole-from-some-parts-0617" target="_new">MIT News</a> / <a href="https://techxplore.com/news/2022-06-vision-technique-3d-2d-images.html" target="_new">TechXplore</a> </div> </td> </tr> <!-- NeurMiPs --> <tr> <td> <img border=0 src="img/neurmips.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>NeurMiPs: Neural Mixture of Planar Experts for View Synthesis</strong> <br> <span class="collaborator">Zhi-Hao Lin, </span> Wei-Chiu Ma, <span class="collaborator">Hao-Yu Max Hsu, </span> <span class="collaborator">Yu-Chiang Frank Wang, </span> <span class="collaborator">Shenlong Wang</span> <br> <a href="" target="_new"><b>CVPR 2022</b></a> / <a href="https://zhihao-lin.github.io/neurmips/" target="_new">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2022/papers/Lin_NeurMiPs_Neural_Mixture_of_Planar_Experts_for_View_Synthesis_CVPR_2022_paper.pdf" target="_new">paper</a> / <a href="https://github.com/zhihao-lin/neurmips" target="_new">code</a> / <a href="https://www.youtube.com/watch?v=PV1dCTWL5Oo" target="_new">video</a> </div> </td> </tr> <!-- BARF --> <tr> <td> <img border=0 src="img/barf.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>BARF: Bundle-Adjusting Neural Radiance Fields </strong> <br> <span class="collaborator">Chen-Hsuan Lin, </span> Wei-Chiu Ma, <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Simon Lucey</span> <br> <a href="" target="_new"><b>ICCV 2021</b></a> / <a href="https://chenhsuanlin.bitbucket.io/bundle-adjusting-NeRF/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2104.06405.pdf" target="_new">arXiv</a> / <a href="https://github.com/chenhsuanlin/bundle-adjusting-NeRF" target="_new">code</a> / <a href="https://www.youtube.com/watch?v=dCmCZs2Hpi0" target="_new">video</a> / <a href="https://read.deeplearning.ai/the-batch/issue-95#3d-scene-synthesis-for-the-real-world" target="_new">news coverage (The Batch: DeepLearning.AI)</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <!-- feedback solver --> <tr> <td> <img border=0 src="img/deep-optimizer-teaser.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Deep Feedback Inverse Problem Solver</strong> <br> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Jiayuan Gu, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Raquel Urtasun</span> <!-- <br><span class="highlight"><b>CVPR 2020</b></span> --> <br> <a href="" target="_new"><b>ECCV 2020</b></a> / <a href="papers/eccv20-deep-optimizer/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2101.07719.pdf" target="_new">arXiv</a> / <a href="papers/eccv20-deep-optimizer/img/deep-feedback-inverse-problem-solver-short.mp4" target="_new">short video (1.5 mins)</a> / <a href="papers/eccv20-deep-optimizer/img/deep-feedback-inverse-problem-solver-long.mp4" target="_new">long video (10 mins)</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <!-- LiDARSim --> <tr> <td> <img border=0 src="img/lidarsim.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>LidarSIM: Realistic LiDAR Simulation by Leveraging the Real World</strong> <br> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Kelvin Wong, Wenyuan Zeng, </span> <span class="collaborator">Bin Yang, </span> <span class="collaborator">Shuhan Tan, </span> <span class="collaborator">Mikita Sazanovich, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <!-- <br><span class="highlight"><b>CVPR 2020</b></span> --> <br> <a href="" target="_new"><b>CVPR 2020</b></a> / <a href="http://openaccess.thecvf.com/content_CVPR_2020/papers/Manivasagam_LiDARsim_Realistic_LiDAR_Simulation_by_Leveraging_the_Real_World_CVPR_2020_paper.pdf" target="_new">paper</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <!-- sparse loc --> <tr> <td> <a href="https://arxiv.org/pdf/1908.03274.pdf"><img border=0 src="img/light-loc.gif" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Exploiting Sparse Semantic HD Maps for Self-Driving Vehicle Localization</strong> <br> Wei-Chiu Ma*, <span class="collaborator">Ignacio Tartavull*, </span> <span class="collaborator">Ioan Andrei Bârsan*, </span> <span class="collaborator">Shenlong Wang*, </span> <span class="collaborator">Min Bai, </span> <span class="collaborator">Gellert Mattyus, </span> <span class="collaborator">Namdar Homayounfar, </span> <span class="collaborator">Shrinidhi K. Lakshmikanth, </span> <span class="collaborator">Andrei Pokrovsky, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>IROS 2019</b></a> / <a href="https://arxiv.org/pdf/1908.03274.pdf" target="_new">arXiv</a> / <a href="https://www.youtube.com/watch?v=-_PvPPr7y28" target="_new">video</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <!-- DRISF --> <tr> <td> <a href="papers/cvpr19-drisf/"><img border=0 src="img/cvpr-drisf.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Deep Rigid Instance Scene Flow</strong> <br> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Rui Hu, </span> <span class="collaborator">Yuwen Xiong, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2019</b></a> / <a href="papers/cvpr19-drisf/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/1904.08913.pdf" target="_new">arXiv</a> / <a href="papers/cvpr19-drisf/paper.pdf" target="_new">paper + supp (uncompressed)</a> / <a href="papers/cvpr19-drisf/gn_iteration.gif" target="_new">GN solver gif</a> <br> <br> Deep structured scene flow model that <span class="highlight"><b>rank 1st</b></span> on <a href="http://www.cvlibs.net/datasets/kitti/eval_scene_flow.php">KITTI Scene Flow Benchmark</a>. <br>Faster than prior art by <span class="highlight"><b>800 times</b></span>. </div> </td> </tr> <!-- Deep Pruner --> <tr> <td> <a href="papers/iccv19-deep-pruner/deeppruner.pdf"><img border=0 src="img/deeppruner.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>DeepPruner: Learning Efficient Stereo Matching via Differentiable PatchMatch</strong> <br> <span class="collaborator">Shivam Duggal, </span> <span class="collaborator">Shenlong Wang, </span> Wei-Chiu Ma, <span class="collaborator">Rui Hu, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>ICCV 2019</b></a> / <a href="https://arxiv.org/pdf/1909.05845.pdf" target="_new">arXiv</a> / <a href="https://github.com/uber-research/DeepPruner/" target="_new">code</a> / <a href="https://github.com/uber-research/DeepPruner/tree/master/DifferentiablePatchMatch" target="_new"> differentiable PatchMatch module</a> <br> <br><span class="highlight"><b>Real-time</b></span> stereo estimation (62 ms) via <span class="highlight"><b>Differentiable PatchMatch</b></span>! </div> </td> </tr> <!-- Continuous Conv --> <tr> <td> <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Wang_Deep_Parametric_Continuous_CVPR_2018_paper.pdf"><img border=0 src="img/cont_conv.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Deep Parametric Continuous Convolutional Neural Networks</strong> <br> <span class="collaborator">Shenlong Wang*, </span> <span class="collaborator">Simon Suo*, </span> Wei-Chiu Ma, <span class="collaborator">Andrei Pokrovsky, </span> <span class="collaborator">and Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2018</b></a> / <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Wang_Deep_Parametric_Continuous_CVPR_2018_paper.pdf" target="_new">paper</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <!-- Sun CNN --> <tr> <td> <a href="https://youtu.be/tOVWptik0qI" target="_blank"><img border=0 src="img/lost2.gif" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Find Your Way by Observing the Sun and Other Semantic Cues</strong> <br>Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Marcus A. Brubaker, </span> <span class="collaborator">Sanja Fidler, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>ICRA 2017</b></a> / <a href="http://arxiv.org/pdf/1606.07415.pdf" target="_new">arXiv</a> / <a href="https://youtu.be/tOVWptik0qI" target="_blank">demo video</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> </script> <script id="pubs_by_date" language="text"> <!-- SVR --> <tr> <td> <img border=0 src="img/svr.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Sparse Voxels Rasterization: Real-time High-fidelity Radiance Field Rendering</strong> <br> <span class="collaborator">Cheng Sun, </span> <span class="collaborator">Jaesung Choe, </span> <span class="collaborator">Charles Loop, </span> Wei-Chiu Ma, <span class="collaborator">Yu-Chiang Frank Wang</span> <br> <a href="" target=""><b>arXiv 2024</b></a> / <a href="https://arxiv.org/pdf/2412.04459" target="">arXiv</a> </div> </td> </tr> <!-- eval3d --> <tr> <td> <img border=0 src="img/eval3d.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Eval3D: Interpretable and Fine-grained Evaluation for 3D Generation</strong> <br> <span class="collaborator">Shivam Duggal*, </span> <span class="collaborator">Yushi Hu*, </span> <span class="collaborator">Oscar Michel, </span> <span class="collaborator">Aniruddha Kembhavi, </span> <span class="collaborator">William T. Freeman, </span> <span class="collaborator">Noah A. Smith, </span> <span class="collaborator">Ranjay Krishna, </span> <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Ali Farhadi, </span> Wei-Chiu Ma <br> <a href="" target=""><b>arXiv 2024</b></a> / <a href="" target="">arXiv</a> </div> </td> </tr> <!-- Odin --> <tr> <td> <img border=0 src="img/odin.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>From an Image to a Scene: Learning to Imagine the World from a Million 360° Videos</strong> <br> <span class="collaborator">Matthew Wallingford, </span> <span class="collaborator">Anand Bhattad, </span> <span class="collaborator">Aditya Kusupati, </span> <span class="collaborator">Vivek Ramanujan, </span> <span class="collaborator">Matt Deitke, </span> <span class="collaborator">Sham Kakade, </span> <span class="collaborator">Aniruddha Kembhavi, </span> <span class="collaborator">Roozbeh Mottaghi, </span> Wei-Chiu Ma, <span class="collaborator">Ali Farhadi</span> <br> <a href="" target=""><b>NeurIPS 2024</b></a> / <a href="https://mattwallingford.github.io/ODIN/" target="">project page</a> / <a href="https://arxiv.org/pdf/2412.07770" target="">arXiv</a> / <a href="https://huggingface.co/datasets/mwallingford/360-1M/tree/main" target="">huggingface</a> </div> </td> </tr> <!-- multilingual --> <tr> <td> <img border=0 src="img/multilingual.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Multilingual Diversity Improves Vision-Language Representations</strong> <br> <span class="collaborator">Thao Nguyen, </span> <span class="collaborator">Matthew Wallingford, </span> <span class="collaborator">Sebastin Santy, </span> Wei-Chiu Ma, <span class="collaborator">Sewoong Oh, </span> <span class="collaborator">Ludwig Schmidt, </span> <span class="collaborator">Pang Wei Koh, </span> <span class="collaborator">Ranjay Krishna</span> <br> <a href="" target=""><b>NeurIPS 2024</b></a> / <a href="https://arxiv.org/pdf/2405.16915" target="">paper</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <!-- task me anything --> <tr> <td> <img border=0 src="img/task-me-anything.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Task Me Anything</strong> <br> <span class="collaborator">Jieyu Zhang, </span> <span class="collaborator">Weikai Huang, </span> <span class="collaborator">Zixian Ma, </span> <span class="collaborator">Oscar Michel, </span> <span class="collaborator">Dong He, </span> <span class="collaborator">Tanmay Gupta, </span> Wei-Chiu Ma, <span class="collaborator">Ali Farhadi, </span> <span class="collaborator">Aniruddha Kembhavi, </span> <span class="collaborator">Ranjay Krishna</span> <br> <a href="" target=""><b>NeurIPS 2024 Datasets and Benchmarks</b></a> / <a href="https://www.task-me-anything.org/" target="">project page</a> / <a href="https://arxiv.org/pdf/2406.11775" target="">paper</a> / <a href="https://huggingface.co/spaces/zixianma/TaskMeAnything-UI" target="">huggingface</a> / <a href="https://github.com/JieyuZ2/TaskMeAnything" target="">code</a> </div> </td> </tr> <!-- BLINK --> <tr> <td> <img border=0 src="img/blink.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>BLINK: Multimodal Large Language Models Can See but Not Perceive</strong> <br> <span class="collaborator">Xingyu Fu*, </span> <span class="collaborator">Yushi Hu*, </span> <span class="collaborator">Bangzheng Li, </span> <span class="collaborator">Yu Feng, </span> <span class="collaborator">Haoyu Wang, </span> <span class="collaborator">Xudong Lin, </span> <span class="collaborator">Dan Roth, </span> <span class="collaborator">Noah A. Smith, </span> Wei-Chiu Ma<sup>&#8224;</sup>, <span class="collaborator">Ranjay Krishna<sup>&#8224;</sup></span> <br> <a href="" target=""><b>ECCV 2024</b></a> / <a href="https://zeyofu.github.io/blink/" target="">project page</a> / <a href="https://arxiv.org/pdf/2404.12390" target="">paper</a> / <a href="https://huggingface.co/datasets/BLINK-Benchmark/BLINK" target="">dataset</a> / <a href="https://eval.ai/web/challenges/challenge-page/2287/overview" target="">Eval AI</a> / <a href="https://github.com/zeyofu/BLINK_Benchmark" target="">code</a> </div> </td> </tr> <!-- Video2Game --> <tr> <td> <img border=0 src="img/video2game.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Video2Game: Real-time, Interactive, Realistic and Browser-Compatible Environment from a Single Video</strong> <br> <span class="collaborator">Hongchi Xia, </span> <span class="collaborator">Zhi-Hao Lin, </span> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang</span> <br> <a href="" target=""><b>CVPR 2024</b></a> / <a href="https://video2game.github.io/" target="">project page</a> / <a href="https://arxiv.org/pdf/2404.09833" target="">paper</a> / <a href="https://video2game.github.io/src/garden/index.html" target="">shooting demo</a> / <a href="https://github.com/video2game/video2game" target="">code</a> </div> </td> </tr> <!-- ExtraNeRF --> <tr> <td> <img border=0 src="img/extra_nerf.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>ExtraNeRF: Visibility-Aware View Extrapolation of Neural Radiance Fields with Diffusion Models</strong> <br> <span class="collaborator">Meng-Li Shih, </span> Wei-Chiu Ma, <span class="collaborator">Lorenzo Boyice, </span> <span class="collaborator">Aleksander Holynski, </span> <span class="collaborator">Forrester Cole, </span> <span class="collaborator">Brian Curless, </span> <span class="collaborator">Janne Kontkanen</span> <br> <a href="" target=""><b>CVPR 2024</b></a> / <a href="https://arxiv.org/pdf/2406.06133" target="">arXiv</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/many-nerf.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Structure from Duplicates: Neural Inverse Graphics from a Pile of Objects</strong> <br> <span class="collaborator">Tianhang Cheng, </span> Wei-Chiu Ma, <span class="collaborator">Kaiyu Guan, </span> <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Shenlong Wang</span> <br> <a href="" target=""><b>NeurIPS 2023</b></a> / <a href="https://tianhang-cheng.github.io/SfD-project.github.io/" target="">project page</a> / <a href="https://arxiv.org/pdf/2401.05236" target=""><b>arXiv</b></a> </div> </td> </tr> <tr> <td> <img border=0 src="img/light-sim.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>LightSim: Neural Lighting Simulation for Urban Scenes </strong> <br> <span class="collaborator">Ava Pun*, </span> <span class="collaborator">Gary Sun*, </span> <span class="collaborator">Jingkang Wang*, </span> <span class="collaborator">Yun Chen, </span> <span class="collaborator">Ze Yang, </span> <span class="collaborator">Sivabalan Manivasagam, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>NeurIPS 2023</b></a> / <a href="https://arxiv.org/pdf/2312.06654" target="">arXiv</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/unisim.mp4" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>UniSim: A Neural Closed-Loop Sensor Simulator</strong> <br> <span class="collaborator">Ze Yang*, </span> <span class="collaborator">Yun Chen*, </span> <span class="collaborator">Jingkang Wang*, </span> <span class="collaborator">Sivabalan Manivasagam*, </span> Wei-Chiu Ma, <span class="collaborator">Joyce Anqi Yang, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CVPR 2023</b></a> / <a href="https://waabi.ai/unisim/" target="">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2023/papers/Yang_UniSim_A_Neural_Closed-Loop_Sensor_Simulator_CVPR_2023_paper.pdf" target="">paper</a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/unisim/unisim_final_v2_4k.mp4" target="">4K demo</a> / <a href="https://waabi.ai/wp-content/uploads/2023/05/UniSim-video_compressed.mp4" target="">video (8 mins)</a> <br> <span class="highlight"><b>Highlight presentation</b></span> </div> </td> </tr> <tr> <td> <img border=0 src="img/lidar-gen.mp4" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>UltraLiDAR: Learning Compact Representations for LiDAR Completion and Generation</strong> <br> <span class="collaborator">Yuwen Xiong, </span> Wei-Chiu Ma, <span class="collaborator">Jingkang Wang, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CVPR 2023</b></a> / <a href="https://waabi.ai/ultralidar/" target="">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2023/papers/Xiong_Learning_Compact_Representations_for_LiDAR_Completion_and_Generation_CVPR_2023_paper.pdf" target="">paper</a> / <a href="https://waabi.ai/wp-content/uploads/2023/05/UltraLidar-video.mov" target="">video (1 min)</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/thermal-im.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>What Happened 3 Seconds Ago? Inferring the Past with Thermal Imaging</strong> <br> <span class="collaborator">Zitian Tang*, </span> <span class="collaborator">Wenjie Yeh*, </span> Wei-Chiu Ma, <span class="collaborator">Hang Zhao</span> <br> <a href="" target=""><b>CVPR 2023</b></a> / <a href="https://arxiv.org/pdf/2304.13651.pdf" target="">arXiv</a> / <a href="https://github.com/ZitianTang/Thermal-IM" target="">dataset</a> </div> </td> </tr> <tr> <td> <img border=0 src="papers/neurips22-sgam/sgam.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>SGAM: Building a Virtual 3D World through Simultaneous Generation and Mapping</strong> <br> <span class="collaborator">Yuan Shen, </span> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang</span> <br> <a href="" target=""><b>NeurIPS 2022</b></a> / <a href="https://yshen47.github.io/sgam/" target="">project page</a> / <a href="https://openreview.net/pdf?id=17KCLTbRymw" target="">paper</a> / <a href="https://github.com/yshen47/SGAM" target="">code</a> </div> </td> </tr> <tr> <td> <img border=0 src="papers/corl22-cad-sim/cad-sim.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>CADSim: Robust and Scalable in-the-wild 3D Reconstruction for Realistic and Controllable Sensor Simulation</strong> <br> <span class="collaborator">Jingkang Wang, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Yun Chen, </span> <span class="collaborator">Ze Yang, </span> <span class="collaborator">Ioan Andrei Bârsan, </span> <span class="collaborator">Joyce Anqi Yang, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target=""><b>CoRL 2022</b></a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/cadsim.html" target="">project page</a> / <a href="https://openreview.net/pdf?id=Mp3Y5jd7rnW" target="">paper</a> / <a href="http://www.cs.toronto.edu/~wangjk/publications/cadsim/cadsim_corl_denoise.mp4" target="">video</a> </div> </td> </tr> <tr> <td> <img border=0 src="papers/corl22-mira/mira.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>MIRA: Mental Imagery for Robotic Affordances</strong> <br> <span class="collaborator">Lin Yen-Chen, </span> <span class="collaborator">Pete Florence, </span> <span class="collaborator">Andy Zeng, </span> <span class="collaborator">Jonathan T. Barron, </span> <span class="collaborator">Yilun Du, </span> Wei-Chiu Ma, <span class="collaborator">Anthony Simeonov, </span> <span class="collaborator">Alberto Rodriguez Garcia, </span> <span class="collaborator">Phillip Isola</span> <br> <a href="" target=""><b>CoRL 2022</b></a> / <a href="http://yenchenlin.me/mira/" target="">project page</a> / <a href="https://arxiv.org/pdf/2212.06088.pdf" target="">paper</a> / <a href="https://www.youtube.com/watch?v=3GH-s9db5e0&ab_channel=MIRA" target="">video</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/vc-3.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Virtual Correspondence: Humans as a Cue for Extreme-View Geometry</strong> <br> Wei-Chiu Ma, <span class="collaborator">Anqi Joyce Yang, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Raquel Urtasun, </span> <span class="collaborator">Antonio Torralba</span> <br> <a href="" target="_new"><b>CVPR 2022</b></a> / <a href="virtual-correspondence" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2206.08365.pdf" target="_new">paper</a> / <a href="https://virtual-correspondence.github.io/img/vc_demo.mp4" target="_new">video (1.5 mins)</a> / <a href="https://youtu.be/W9odd2F2Bx4" target="_new">video (5 mins)</a> / <a href="https://news.mit.edu/2022/seeing-whole-from-some-parts-0617" target="_new">MIT News</a> / <a href="https://techxplore.com/news/2022-06-vision-technique-3d-2d-images.html" target="_new">TechXplore</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/neurmips.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>NeurMiPs: Neural Mixture of Planar Experts for View Synthesis</strong> <br> <span class="collaborator">Zhi-Hao Lin, </span> Wei-Chiu Ma, <span class="collaborator">Hao-Yu Max Hsu, </span> <span class="collaborator">Yu-Chiang Frank Wang, </span> <span class="collaborator">Shenlong Wang</span> <br> <a href="" target="_new"><b>CVPR 2022</b></a> / <a href="https://zhihao-lin.github.io/neurmips/" target="_new">project page</a> / <a href="https://openaccess.thecvf.com/content/CVPR2022/papers/Lin_NeurMiPs_Neural_Mixture_of_Planar_Experts_for_View_Synthesis_CVPR_2022_paper.pdf" target="_new">paper</a> / <a href="https://github.com/zhihao-lin/neurmips" target="_new">code</a> / <a href="https://www.youtube.com/watch?v=PV1dCTWL5Oo" target="_new">video</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/sdf_secrets.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Mending Neural Implicit Modeling for 3D Vehicle Reconstruction in the Wild</strong> <br> <span class="collaborator">Shivam Duggal*, </span> <span class="collaborator">Zihao Wang*, </span> Wei-Chiu Ma, <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Justin Liang, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target="_new"><b>WACV 2022</b></a> / <a href="https://arxiv.org/pdf/2101.06860.pdf" target="_new">arXiv</a> / <a href="https://www.youtube.com/watch?v=IRygme5J-Ng" target="_new">video</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/barf.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>BARF: Bundle-Adjusting Neural Radiance Fields </strong> <br> <span class="collaborator">Chen-Hsuan Lin, </span> Wei-Chiu Ma, <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Simon Lucey</span> <br> <a href="" target="_new"><b>ICCV 2021</b></a> / <a href="https://chenhsuanlin.bitbucket.io/bundle-adjusting-NeRF/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2104.06405.pdf" target="_new">arXiv</a> / <a href="https://github.com/chenhsuanlin/bundle-adjusting-NeRF" target="_new">code</a> / <a href="https://www.youtube.com/watch?v=dCmCZs2Hpi0" target="_new">video</a> / <a href="https://read.deeplearning.ai/the-batch/issue-95#3d-scene-synthesis-for-the-real-world" target="_new">news coverage (The Batch: DeepLearning.AI)</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <tr> <td> <img border=0 src="img/s3-skinning.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>S3: Neural Shape, Skeleton, and Skinning Fields for 3D Human Modeling </strong> <br> <span class="collaborator">Ze Yang, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Zeng Huang, </span> Wei-Chiu Ma, <span class="collaborator">Xinchen Yan, </span> <span class="collaborator">Ersin Yumer, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target="_new"><b>CVPR 2021</b></a> / <a href="https://arxiv.org/pdf/2101.06571.pdf" target="_new">arXiv</a> / <a href="https://www.youtube.com/watch?v=oCpFJZrVDCs" target="_new">video (5 mins)</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/lime.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Recovering and Simulating Pedestrians in the Wild </strong> <br> <span class="collaborator">Ze Yang, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Ming Liang, </span> <span class="collaborator">Bin Yang, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target="_new"><b>CoRL 2020</b></a> / <a href="https://arxiv.org/abs/2011.08106" target="_new">arXiv</a> / <a href="https://www.youtube.com/watch?v=iBsIoCJ9I1s" target="_new">video</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <tr> <td> <img border=0 src="img/deep-optimizer-teaser.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Deep Feedback Inverse Problem Solver</strong> <br> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Jiayuan Gu, </span> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Antonio Torralba, </span> <span class="collaborator">Raquel Urtasun</span> <!-- <br><span class="highlight"><b>CVPR 2020</b></span> --> <br> <a href="" target="_new"><b>ECCV 2020</b></a> / <a href="papers/eccv20-deep-optimizer/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/2101.07719.pdf" target="_new">arXiv</a> / <a href="papers/eccv20-deep-optimizer/img/deep-feedback-inverse-problem-solver-short.mp4" target="_new">short video (1.5 mins)</a> / <a href="papers/eccv20-deep-optimizer/img/deep-feedback-inverse-problem-solver-long.mp4" target="_new">long video (10 mins)</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <tr> <td> <img border=0 src="img/shape-comp.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Weakly-supervised 3D Shape Completion in the Wild</strong> <br> <span class="collaborator">Jiayuan Gu, </span> Wei-Chiu Ma, <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Wenyuan Zeng, </span> <span class="collaborator">Zihao Wang, </span> <span class="collaborator">Yuwen Xiong, </span> <span class="collaborator">Hao Su, </span> <span class="collaborator">Raquel Urtasun</span> <!-- <br><span class="highlight"><b>CVPR 2020</b></span> --> <br> <a href="" target="_new"><b>ECCV 2020</b></a> / <a href="https://arxiv.org/pdf/2008.09110.pdf" target="_new">arXiv</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <tr> <td> <img border=0 src="img/levelset-rcnn.png" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>LevelSet R-CNN: A Deep Variational Method for Instance Segmentation</strong> <br> <span class="collaborator">Namdar Homayounfar*, </span> <span class="collaborator">Yuwen Xiong*, </span> <span class="collaborator">Justin Liang*, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target="_new"><b>ECCV 2020</b></a> / <a href="https://arxiv.org/pdf/2007.15629.pdf" target="_new">arXiv</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/video-compression.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>Conditional Entropy Coding for Efficient Video Compression</strong> <br> <span class="collaborator">Jerry Junkai Liu, </span> <span class="collaborator">Shenlong Wang, </span> Wei-Chiu Ma, <span class="collaborator">Meet Shah, </span> <span class="collaborator">Rui Hu, </span> <span class="collaborator">Pranaab Dhawan, </span> <span class="collaborator">Raquel Urtasun</span> <br> <a href="" target="_new"><b>ECCV 2020</b></a> / <a href="https://arxiv.org/pdf/2008.09180.pdf" target="_new">arXiv</a> </div> </td> </tr> <tr> <td> <a href="https://arxiv.org/pdf/1912.02801.pdf"><img border=0 src="img/polytransform-new" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>PolyTransform: Deep Polygon Transformer for Instance Segmentation </strong> <br> <span class="collaborator">Justin Liang, </span> <span class="collaborator">Namdar Homayounfar, </span> Wei-Chiu Ma, <span class="collaborator">Yuwen Xiong, </span> <span class="collaborator">Rui Hu, </span> <span class="collaborator">Raquel Urtasun </span> <br><a href="" target="_new"><b>CVPR 2020</b></a> / <a href="https://arxiv.org/pdf/1912.02801.pdf" target="_new">arXiv</a> / <a href="http://justin-liang.com/papers/Liang_PolyTransform_Supp.pdf" target="_new">supp</a> / <a href="http://justin-liang.com/papers/Liang_PolyTransform_Video.zip" target="_new">video</a> / <a href="" target="_new">code (coming soon!)</a> <br> <br>State-of-the-art performance on Cityscapes! Consistent improvements across all backbones! </div> </td> </tr> <tr> <td> <img border=0 src="img/lidarsim.gif" class="publogo"> </td> <td> <div class="publication"> <br> <p><strong>LidarSIM: Realistic LiDAR Simulation by Leveraging the Real World</strong> <br> <span class="collaborator">Sivabalan Manivasagam, </span> <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Kelvin Wong, Wenyuan Zeng, </span> <span class="collaborator">Bin Yang, </span> <span class="collaborator">Shuhan Tan, </span> <span class="collaborator">Mikita Sazanovich, </span> Wei-Chiu Ma, <span class="collaborator">Raquel Urtasun</span> <!-- <br><span class="highlight"><b>CVPR 2020</b></span> --> <br> <a href="" target="_new"><b>CVPR 2020</b></a> / <a href="http://openaccess.thecvf.com/content_CVPR_2020/papers/Manivasagam_LiDARsim_Realistic_LiDAR_Simulation_by_Leveraging_the_Real_World_CVPR_2020_paper.pdf" target="_new">paper</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <tr> <td> <a href="https://arxiv.org/pdf/1908.03274.pdf"><img border=0 src="img/light-loc.gif" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Exploiting Sparse Semantic HD Maps for Self-Driving Vehicle Localization</strong> <br> Wei-Chiu Ma*, <span class="collaborator">Ignacio Tartavull*, </span> <span class="collaborator">Ioan Andrei Bârsan*, </span> <span class="collaborator">Shenlong Wang*, </span> <span class="collaborator">Min Bai, </span> <span class="collaborator">Gellert Mattyus, </span> <span class="collaborator">Namdar Homayounfar, </span> <span class="collaborator">Shrinidhi K. Lakshmikanth, </span> <span class="collaborator">Andrei Pokrovsky, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>IROS 2019</b></a> / <a href="https://arxiv.org/pdf/1908.03274.pdf" target="_new">arXiv</a> / <a href="https://www.youtube.com/watch?v=-_PvPPr7y28" target="_new">video</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <tr> <td> <a href="papers/cvpr19-drisf/"><img border=0 src="img/cvpr-drisf.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Deep Rigid Instance Scene Flow</strong> <br> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Rui Hu, </span> <span class="collaborator">Yuwen Xiong, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2019</b></a> / <a href="papers/cvpr19-drisf/" target="_new">project page</a> / <a href="https://arxiv.org/pdf/1904.08913.pdf" target="_new">arXiv</a> / <a href="papers/cvpr19-drisf/paper.pdf" target="_new">paper + supp (uncompressed)</a> / <a href="papers/cvpr19-drisf/gn_iteration.gif" target="_new">GN solver gif</a> <br> <br> Deep structured scene flow model that <span class="highlight"><b>rank 1st</b></span> on <a href="http://www.cvlibs.net/datasets/kitti/eval_scene_flow.php">KITTI Scene Flow Benchmark</a>. <br>Faster than prior art by <span class="highlight"><b>800 times</b></span>. </div> </td> </tr> <tr> <td> <a href=""><img border=0 src="img/cvpr-road-boundary.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Convolutional Recurrent Network for Road Boundary Extraction</strong> <br> <span class="collaborator">Justin Liang*, </span> <span class="collaborator">Namdar Homayounfar*, </span> Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2019</b></a> / <a href="https://nhoma.github.io/papers/road_cvpr19.pdf" target="_new">paper</a> / <a href="https://nhoma.github.io/papers/road_cvpr19_supp.pdf" target="_new">supp</a> </div> </td> </tr> <tr> <td> <a href="papers/iccv19-deep-pruner/deeppruner.pdf"><img border=0 src="img/deeppruner.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>DeepPruner: Learning Efficient Stereo Matching via Differentiable PatchMatch</strong> <br> <span class="collaborator">Shivam Duggal, </span> <span class="collaborator">Shenlong Wang, </span> Wei-Chiu Ma, <span class="collaborator">Rui Hu, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>ICCV 2019</b></a> / <a href="https://arxiv.org/pdf/1909.05845.pdf" target="_new">arXiv</a> / <a href="https://github.com/uber-research/DeepPruner/" target="_new">code</a> / <a href="https://github.com/uber-research/DeepPruner/tree/master/DifferentiablePatchMatch" target="_new"> differentiable PatchMatch module</a> <br> <br><span class="highlight"><b>Real-time</b></span> stereo estimation (62 ms) via <span class="highlight"><b>Differentiable PatchMatch</b></span>! </div> </td> </tr> <tr> <td> <a href="http://www.mit.edu/~hangzhao/videos/SoM_supp.mp4"><img border=0 src="img/sound_of_motions" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>The Sound of Motions</strong> <br> <span class="collaborator">Hang Zhao, </span> <span class="collaborator">Chuang Gan, </span> Wei-Chiu Ma, <span class="collaborator">Antonio Torralba </span> <br> <a href="" target="_new"><b>ICCV 2019</b></a> / <a href="https://arxiv.org/pdf/1904.05979.pdf" target="_new">arXiv</a> / <a href="http://www.mit.edu/~hangzhao/videos/SoM_supp.mp4" target="_new">demon video</a> </div> </td> </tr> <tr> <td> <a href=""><img border=0 src="img/iccv19-dag.jpg" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>DAG-Mapper: Learning to Map by Discovering Lane Topology</strong> <br> <span class="collaborator">Namdar Homayounfar, </span> Wei-Chiu Ma, <span class="collaborator">Justin Liang, </span> <span class="collaborator">Xinyu Wu, </span> <span class="collaborator">Jack Fan, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>ICCV 2019</b></a> / <a href="https://nhoma.github.io/papers/dagmapper_iccv19.pdf" target="_new">paper</a> / <a href="https://nhoma.github.io/papers/dagmapper_iccv19_supp.pdf" target="_new">supp</a> </div> </td> </tr> <tr> <td> <a href="papers/eccv18-intrinsics/top.pdf"><img border=0 src="img/cover_intrinsic.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Single Image Intrinsic Decomposition without a Single Intrinsic Image</strong> <br> Wei-Chiu Ma, <span class="collaborator">Hang Chu, </span> <span class="collaborator">Bolei Zhou, </span> <span class="collaborator">Raquel Urtasun, </span> <span class="collaborator">Antonio Torralba </span> <br> <a href="" target="_new"><b>ECCV 2018</b></a> / <a href="papers/eccv18-intrinsics/top.pdf" target="_new">paper</a> </div> </td> </tr> <tr> <td> <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Wang_Deep_Parametric_Continuous_CVPR_2018_paper.pdf"><img border=0 src="img/cont_conv.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Deep Parametric Continuous Convolutional Neural Networks</strong> <br> <span class="collaborator">Shenlong Wang*, </span> <span class="collaborator">Simon Suo*, </span> Wei-Chiu Ma, <span class="collaborator">Andrei Pokrovsky, </span> <span class="collaborator">and Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2018</b></a> / <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Wang_Deep_Parametric_Continuous_CVPR_2018_paper.pdf" target="_new">paper</a> <br> <span class="highlight"><b>Spotlight presentation</b></span> </div> </td> </tr> <tr> <td> <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Chu_SurfConv_Bridging_3D_CVPR_2018_paper.pdf"><img border=0 src="img/surf_conv.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>SurfConv: Bridging 3D and 2D Convolution for RGBD Images</strong> <br> <span class="collaborator">Hang Chu, </span> Wei-Chiu Ma, <span class="collaborator">Kaustav Kundu, </span> <span class="collaborator">Raquel Urtasun, </span> <span class="collaborator">and Sanja Fidler </span> <br> <a href="" target="_new"><b>CVPR 2018</b></a> / <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Chu_SurfConv_Bridging_3D_CVPR_2018_paper.pdf" target="_new">paper</a> / <a href="https://github.com/chuhang/SurfConv" target="_new">code</a> </div> </td> </tr> <tr> <td> <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Homayounfar_Hierarchical_Recurrent_Attention_CVPR_2018_paper.pdf"><img border=0 src="img/snake.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Hierarchical Recurrent Attention Networks for Structured Online Maps</strong> <br> <span class="collaborator">Namdar Homayounfar, </span> Wei-Chiu Ma, <span class="collaborator">Shrinidhi K. Lakshmikanth, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>CVPR 2018</b></a> / <a href="http://openaccess.thecvf.com/content_cvpr_2018/papers/Homayounfar_Hierarchical_Recurrent_Attention_CVPR_2018_paper.pdf" target="_new">paper</a> / <a href="https://nhoma.github.io/papers/hran_cvpr18_supp.pdf" target="_new">supp</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/book1.png" class="publogo"> </td> <td> <div class="publication"> <p><strong>Activity Forecasting: An Invitation to Predictive Perception</strong> <br> <span class="collaborator">Kris M. Kitani, </span> <span class="collaborator">De-An Huang, </span> Wei-Chiu Ma <br> <a href="" target="_new"><b>Group and Crowd Behavior for Computer Vision. Chapter 12, 2017</b></a> / <a href="http://www.sciencedirect.com/science/article/pii/B978012809276700014X" target="_new">link</a> </div> </td> </tr> <tr> <td> <a href="https://youtu.be/tOVWptik0qI" target="_blank"><img border=0 src="img/lost2.gif" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Find Your Way by Observing the Sun and Other Semantic Cues</strong> <br>Wei-Chiu Ma, <span class="collaborator">Shenlong Wang, </span> <span class="collaborator">Marcus A. Brubaker, </span> <span class="collaborator">Sanja Fidler, </span> <span class="collaborator">Raquel Urtasun </span> <br> <a href="" target="_new"><b>ICRA 2017</b></a> / <a href="http://arxiv.org/pdf/1606.07415.pdf" target="_new">arXiv</a> / <a href="https://youtu.be/tOVWptik0qI" target="_blank">demo video</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <tr> <td> <a href="http://arxiv.org/pdf/1604.01431.pdf"><img border=0 src="img/FPIOC.gif" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>Forecasting Interactive Dynamics of Pedestrians with Fictitious Play</strong> <br> Wei-Chiu Ma, <span class="collaborator">De-An Huang, </span> <span class="collaborator">Namhoon Lee, </span> <span class="collaborator">Kris M. Kitani </span> <br> <a href="" target="_new"><b>CVPR 2017</b></a> / <a href="http://arxiv.org/pdf/1604.01431.pdf" target="_new">arXiv</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/CVPR2015.png" class="publogo"> </td> <td> <div class="publication"> <p><strong>How Do We Use Our Hands? Discovering a Diverse Set of Common Grasps</strong> <br> <span class="collaborator">De-An Huang, </span> Wei-Chiu. Ma*, <span class="collaborator">Minghuan Ma*, </span> <span class="collaborator">K. M. Kitani </span> <br> <a href="" target="_new"><b>CVPR 2015</b></a> / <a href="https://scholar.google.com/citations?view_op=view_citation&hl=en&user=yv3sH74AAAAJ&cstart=20&sortby=pubdate&citation_for_view=yv3sH74AAAAJ:HeT0ZceujKMC" target="_new">paper</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/Hand-Object-Interaction.png" class="publogo"> </td> <td> <div class="publication"> <p><strong>Recognizing Hand-Object Interactions in Wearable Camera Videos</strong> <br> <span class="collaborator">Tatsuya Ishihara, </span> <span class="collaborator">Kris M. Kitani, </span> Wei-Chiu Ma, <span class="collaborator">Hironobu Takagi, </span> <span class="collaborator">Chieko Asakawa </span> <br> <a href="" target="_new"><b>ICIP 2015</b></a> / <a href="http://www.cs.cmu.edu/~kkitani/pdf/IKMTA-ICIP2015.pdf" target="_new">paper</a> </div> </td> </tr> <tr> <td> <img border=0 src="img/TrafficGA.png" class="publogo"> </td> <td> <div class="publication"> <p><strong>Novel traffic signal timing adjustment strategy based on Genetic Algorithm</strong> <br> <span class="collaborator">Hsiao-Yu Tung*, </span> Wei-Chiu Ma*, <span class="collaborator">Tian-Li Yu </span> <br> <a href="" target="_new"><b>CEC 2014</b></a> / <a href="http://www.andrew.cmu.edu/user/weichium/publications/TrafficGA.pdf" target="_new">paper</a> <br> <span class="highlight"><b>Oral presentation</b></span> </div> </td> </tr> <tr> <td> <a href="https://www.youtube.com/embed/TUwk1DSgl1Q"><img border=0 src="img/TDTOS-concept.png" class="publogo"></a> </td> <td> <div class="publication"> <p><strong>TDTOS: T-Shirt Design and Try On System</strong> <br> <span class="collaborator">Chen-Yu Hsu*, </span> <span class="collaborator">Chi-Hsien Yen*, </span> Wei-Chiu Ma*, <span class="collaborator">Shao-Yi Chien </span> <br> <a href="" target="_new"><b>Asia-Pacific Workshop on FPGA Applications 2012</b></a> / <!-- <br><b>Oral Presentation</b> (accpetance rate: 4.8%) --> <a href="https://www.youtube.com/embed/TUwk1DSgl1Q" target="_new">demo video</a> <br> <span class="highlight"><b>Oral presentation</b></span> <br> <span class="highlight"><b>Best Application Award</b></span> </div> </td> </tr> </script> </table> </div> </td> </tr> <tr> <td colspan = 7> <p>&nbsp;</p> <h2 id="pro-bono">Pro bono office hour</h2> <p>Inspired by <a href="https://kyunghyuncho.me/office-hour-request/">Prof. Kyunghyun Cho</a> and <a href="https://krrish94.github.io/">Krishna Murthy</a>, starting January 2021, I have decided to commit 1~2 hours every week to provide guidance, suggestions, and/or mentorships for students from underrepresented groups or whoever is in need. Please fill in this <a href="https://docs.google.com/forms/d/1uYA95BMyFjnvY2-XDbwSAlTz79Ip55cytzrNoROS9ko/edit">form</a> if you are interested.</p> Need more (diverse) opinions? Consider talking to people with different expertise or from different background: <a href="https://www.tongzhouwang.info">Tongzhou Wang (ML/RL)</a>, <a href="https://alexander-h-liu.github.io">Alexander Haojan Liu (Speech/NLP)</a>, <a href="https://zhijianliu.com">Zhijian Liu (ML)</a>, <a href="https://www.zyrianov.org">Vlas Zyrianov (CV)</a>, <a href="https://yshen47.github.io">Yuan Shen (CV)</a>, <a href="https://www.cs.toronto.edu/~jungao/#pro-bono">Jun Gao (CV/Graphics)</a>. </td> </tr> <tr> <td colspan = 7> <h2 id="misc">Misc.</h2> <ul> <li> I enjoy playing/watching all kinds of sports, including but not limited to badminton, basketball, baseball, tennis, running, etc. My favorite sports players and my favorite games (click their names!) are: <ul> <li><a href="https://www.youtube.com/watch?v=GTJwoWHMEw0">Kobe Bryant</a> (basketball) </li> <li> <a href="https://www.youtube.com/watch?v=rctOtFOXco8">Roger Federer</a> (tennis)</li> <li> <a href="https://www.youtube.com/watch?v=Hf-n1yfd8II">Lee Chong Wei</a> (badminton)</li> </ul> </li> <br> <li> I like chinese pop music and my favorite singer is <a href="https://en.wikipedia.org/wiki/Jay_Chou">Jay Chou</a>. <!-- Here are a few personal favorite songs: --> </li> <br> <li> A few personal favorite papers/books/courses can be found <a href="misc/favorite-books-papers.html">here</a> (unordered and unfinished).</li> </ul> </td> </tr> <!-- <tr> <td colspan = 7> <p>&nbsp;</p> <h2>Collaborators</h2> <ul> <li>Over the past few years, I have had the fortune to work with a bunch of great people: <a href="http://www.cs.toronto.edu/~slwang/">Shenlong Wang</a> (UofT), <a href="http://www.cs.toronto.edu/~mbrubake/">Marcus A. Brubaker</a> (UofT)</li> </ul> </td> </tr> --> </table> </center> </div> </div> <br><center><a href="http://accessibility.mit.edu/">Accessibility</a></center> </center> <!-- Start of StatCounter Code for Default Guide --> <script type="text/javascript"> var sc_project=11104199; var sc_invisible=1; var sc_security="50dd2089"; var scJsHost = (("https:" == document.location.protocol) ? "https://secure." : "http://www."); document.write("<sc"+"ript type='text/javascript' src='" + scJsHost+ "statcounter.com/counter/counter.js'></"+"script>"); </script> <noscript><div class="statcounter"><a title="free hit counter" href="http://statcounter.com/" target="_blank"><img class="statcounter" src="//c.statcounter.com/11104199/0/50dd2089/1/" alt="free hit counter"></a></div></noscript> <!-- End of StatCounter Code for Default Guide --> <script>showPubs(0);</script> </body> </html>

Pages: 1 2 3 4 5 6 7 8 9 10