<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Localization on ViCoS Lab</title>
    <link>/tags/localization/</link>
    <description>Recent content in Localization on ViCoS Lab</description>
    <generator>Hugo</generator>
    <language>en-us</language>
    <atom:link href="/tags/localization/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Center Direction Network for Grasping Point Localization on Cloths</title>
      <link>/publications/tabernik2024center/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>/publications/tabernik2024center/</guid>
      <description>&lt;p&gt;Object grasping is a fundamental challenge in robotics and computer vision, critical for advancing robotic manipulation capabilities. Deformable objects, like fabrics and cloths, pose additional challenges due to their non-rigid nature. In this work, we introduce CeDiRNet-3DoF, a deep-learning model for grasp point detection, with a particular focus on cloth objects. CeDiRNet-3DoF employs center direction regression alongside a localization network, attaining first place in the perception task of ICRA 2023&amp;rsquo;s Cloth Manipulation Challenge. Recognizing the lack of standardized benchmarks in the literature that hinder effective method comparison, we present the ViCoS {Towel} Dataset. This extensive benchmark dataset comprises 8,000 real and 12,000 synthetic images, serving as a robust resource for training and evaluating contemporary data-driven deep-learning approaches. Extensive evaluation revealed CeDiRNet-3DoF&amp;rsquo;s robustness in real-world performance, outperforming state-of-the-art methods, including the latest transformer-based models. Our work bridges a crucial gap, offering a robust solution and benchmark for cloth grasping in computer vision and robotics. Code and dataset are available at: &lt;a href=&#34;https://github.com/vicoslab/CeDiRNet-3DoF&#34;&gt;https://github.com/vicoslab/CeDiRNet-3DoF&lt;/a&gt;&lt;/p&gt;</description>
    </item>
    <item>
      <title>Non-sequential Multi-view Detection, Localization and Identification of People Using Multi-modal Feature Maps</title>
      <link>/publications/mandeljc2012non-sequential/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>/publications/mandeljc2012non-sequential/</guid>
      <description></description>
    </item>
    <item>
      <title>Tracking by Identification Using Computer Vision and Radio</title>
      <link>/publications/mandeljc2013tracking/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>/publications/mandeljc2013tracking/</guid>
      <description>&lt;p&gt;We present a novel system for detection, localization and tracking of multiple people, which fuses a multi-view computer vision approach with a radio-based localization system. The proposed fusion combines the best of both worlds, excellent computer-vision-based localization, and strong identity information provided by the radio system, and is therefore able to perform tracking by identification, which makes it impervious to propagated identity switches. We present comprehensive methodology for evaluation of systems that perform person localization in world coordinate system and use it to evaluate the proposed system as well as its components. Experimental results on a challenging indoor dataset, which involves multiple people walking around a realistically cluttered room, confirm that proposed fusion of both systems significantly outperforms its individual components. Compared to the radio-based system, it achieves better localization results, while at the same time it successfully prevents propagation of identity switches that occur in pure computer-vision-based tracking.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
