<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Qilin-Zhang | CAMCV</title><link>https://camcv.github.io/author/qilin-zhang/</link><atom:link href="https://camcv.github.io/author/qilin-zhang/index.xml" rel="self" type="application/rss+xml"/><description>Qilin-Zhang</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Wed, 01 Jan 2025 00:00:00 +0000</lastBuildDate><image><url>https://camcv.github.io/media/icon_hu_642791a3c5816746.png</url><title>Qilin-Zhang</title><link>https://camcv.github.io/author/qilin-zhang/</link></image><item><title>Computer Vision</title><link>https://camcv.github.io/projects/cv/</link><pubDate>Wed, 01 Jan 2025 00:00:00 +0000</pubDate><guid>https://camcv.github.io/projects/cv/</guid><description>&lt;p>We are the Computer Vision Working Group of CV4DT. Click for more!&lt;/p>
&lt;p>We are the Computer Vision Working Group of CV4DT: (&lt;a href="https://camcv.github.io/author/dr-olaf-wysocki/">Olaf Wysocki&lt;/a>, &lt;a href="https://camcv.github.io/author/haibing-wu/">Haibing Wu&lt;/a>, &lt;a href="https://camcv.github.io/author/qilin-zhang/">Qilin Zhang&lt;/a>, &lt;a href="https://camcv.github.io/author/daniel-lehmberg/">Daniel Lehmberg&lt;/a>, &lt;a href="https://camcv.github.io/author/wanru-yang/">Wanru Yang&lt;/a>).&lt;/p>
&lt;h1 id="research-projects">Research Projects&lt;/h1>
&lt;p>Our projects are primarily &lt;strong>research-oriented&lt;/strong>, aiming for publication in top-tier computer vision venues such as &lt;strong>CVPR&lt;/strong>, &lt;strong>ECCV&lt;/strong> &lt;strong>NeurIPS&lt;/strong>, and similar.&lt;br>
Below is an overview of our ongoing and upcoming research directions.&lt;/p>
&lt;hr>
&lt;h3 id="-structured-3d-object-reconstruction">🏠 Structured 3D Object Reconstruction&lt;/h3>
&lt;p>We aim to reconstruct &lt;strong>structured 3D models&lt;/strong> aligned with interpretable geometric and semantic representations.&lt;br>
This direction builds upon our prior work:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://openaccess.thecvf.com/content/CVPR2025W/USM3D/papers/Tang_Texture2LoD3_Enabling_LoD3_Building_Reconstruction_With_Panoramic_Images_CVPRW_2025_paper.pdf" target="_blank" rel="noopener">&lt;em>Texture2LoD3: Enabling LoD3 Building Reconstruction With Panoramic Images&lt;/em> (CVPR25)&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://openaccess.thecvf.com/content/CVPR2023W/PCV/papers/Wysocki_Scan2LoD3_Reconstructing_Semantic_3D_Building_Models_at_LoD3_Using_Ray_CVPRW_2023_paper.pdf" target="_blank" rel="noopener">&lt;em>Scan2LoD3: Reconstructing semantic 3D building models at LoD3 using ray casting and Bayesian networks&lt;/em> (CVPR23)&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Several project proposals are currently under review to expand this line of research.&lt;/p>
&lt;hr>
&lt;h3 id="-revisiting-geometric-features-for-3d-scene-understanding">🧩 Revisiting Geometric Features for 3D Scene Understanding&lt;/h3>
&lt;p>We revisit &lt;strong>geometric descriptors&lt;/strong> for large-scale &lt;strong>3D semantic segmentation&lt;/strong>, &lt;strong>SSL&lt;/strong>, &lt;strong>3D instance segmentation&lt;/strong>, &lt;strong>3D object pose estimation&lt;/strong>, &lt;strong>3D shape completion&lt;/strong>, studying how handcrafted and learned geometric features can be combined to achieve better generalization across domains. Preliminary findings are available in &lt;a href="https://arxiv.org/pdf/2402.06506" target="_blank" rel="noopener">&lt;em>arXiv:2402.06506&lt;/em>&lt;/a>&lt;/p>
&lt;p>Papers are in the making to expand this line of research.&lt;/p>
&lt;hr>
&lt;h3 id="-sim2real-3d-domain-gap">🏁 Sim2Real 3D Domain Gap&lt;/h3>
&lt;p>We still observe large domain gaps between simulated and real-world data, hampering application of simulated data into real-world challenges and many downstream tasks. We believe in the power of &lt;em>diffusion models&lt;/em> to cater for this gap. Preliminary results, building a framework for running simulations within the unique real-world city twin are here: &lt;a href="https://arxiv.org/abs/2505.17959" target="_blank" rel="noopener">&lt;em>arXiv:2505.17959&lt;/em>&lt;/a>&lt;/p>
&lt;p>One paper is under review, while another draft is in preparation.&lt;/p>
&lt;hr>
&lt;h3 id="-6dof-estimation-using-structured-3d-models">🧭 6DoF Estimation Using Structured 3D Models&lt;/h3>
&lt;p>We explore &lt;strong>structured 3D model representations&lt;/strong> for &lt;strong>6-degree-of-freedom (6DoF) pose estimation&lt;/strong>, targeting improved robustness and interpretability compared to implicit or point-based methods.&lt;br>
This direction builds on related work of &lt;a href="https://proceedings.neurips.cc/paper_files/paper/2024/file/d78ece6613953f46501b958b7bb4582f-Paper-Conference.pdf" target="_blank" rel="noopener">&lt;em>LoD-Loc: Aerial Visual Localization using LoD 3D
Map with Neural Wireframe Alignment&lt;/em> (NeurIPS24)&lt;/a>.&lt;/p>
&lt;p>A new iteration of this work is in preparation for upcoming major conference deadlines.&lt;/p>
&lt;hr>
&lt;h3 id="-geometry-prior-guided-3d-gaussian-splatting">🌌 Geometry-Prior-Guided 3D Gaussian Splatting&lt;/h3>
&lt;p>This project investigates the integration of &lt;strong>geometry-aware priors&lt;/strong> into &lt;strong>3D Gaussian Splatting&lt;/strong> to enhance reconstruction quality and geometric fidelity.&lt;br>
Preliminary findings are available in &lt;a href="https://arxiv.org/pdf/2508.07355" target="_blank" rel="noopener">&lt;em>arXiv:2508.07355&lt;/em>&lt;/a>, and ongoing work extends the framework beyond building-specific scenarios toward &lt;strong>general-purpose 3D environments&lt;/strong>.&lt;/p>
&lt;hr>
&lt;h3 id="-quantifying-uncertainty-of-x">📈 Quantifying Uncertainty of X&lt;/h3>
&lt;p>In this research direction, we explore quantification of uncertainty in various modalities and downstream tasks: From data acqusition, through segmentation to inference. Our rationale is often grounded in Bayesian modeling of uncertainty (but not limited to!). Previously published papers, e.g., on reconstruction uncertainty
&lt;a href="https://openaccess.thecvf.com/content/CVPR2023W/PCV/papers/Wysocki_Scan2LoD3_Reconstructing_Semantic_3D_Building_Models_at_LoD3_Using_Ray_CVPRW_2023_paper.pdf" target="_blank" rel="noopener">&lt;em>Scan2LoD3: Reconstructing semantic 3D building models at LoD3 using ray casting and Bayesian networks&lt;/em> (CVPR23)&lt;/a>.
Currently, we are involved in the funded project, &lt;a href="https://www.asg.ed.tum.de/en/gds/forschung-research/projects/nerf2bim/" target="_blank" rel="noopener">NeRF2BIM&lt;/a>, together with Profs Petzold, Holst and Niessner, where we analyze laser scanning uncertainty and its influence on final 3D object reconstruction.&lt;/p>
&lt;hr>
&lt;h3 id="-dataset-development">🗂️ Dataset Development&lt;/h3>
&lt;p>We also curate and release datasets supporting our main research directions, including:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Facade Segmentation Dataset&lt;/strong> – for large-scale semantic façade parsing; This building upon our worldwide-largest facade dataset &lt;a href="https://openaccess.thecvf.com/content/WACV2025/html/Wysocki_ZAHA_Introducing_the_Level_of_Facade_Generalization_and_the_Large-Scale_WACV_2025_paper.html" target="_blank" rel="noopener">&lt;em>ZAHA&lt;/em> (WACV25)&lt;/a>&lt;/li>
&lt;li>&lt;strong>Point Cloud Completion Dataset&lt;/strong> – for partial-to-complete reconstruction learning&lt;/li>
&lt;li>&lt;strong>3D Object Reconstruction Dataset&lt;/strong> – for structured geometry prediction and analysis&lt;/li>
&lt;/ul>
&lt;p>These datasets promote &lt;strong>reproducible, data-rich 3D research&lt;/strong> across geometry, perception, and robotics.&lt;/p>
&lt;p>We are always looking for fantastic persons to join us for the following project collaboration!&lt;/p></description></item></channel></rss>