<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>DSpace Comunidad :</title>
    <link>https://hdl.handle.net/11000/418</link>
    <description />
    <pubDate>Wed, 08 Apr 2026 10:44:25 GMT</pubDate>
    <dc:date>2026-04-08T10:44:25Z</dc:date>
    <item>
      <title>Study of Qu-Based Resonant Microwave Sensors and Design of 3-D-Printed Devices Dedicated to Glucose Monitoring</title>
      <link>https://hdl.handle.net/11000/38844</link>
      <description>Título : Study of Qu-Based Resonant Microwave Sensors and Design of 3-D-Printed Devices Dedicated to Glucose Monitoring
Autor : Juan, Carlos G.; Potelon, Benjamin; Quendo, Cédric; García-Martínez, Héctor; Ávila-Navarro, Ernesto; Bronchalo, Enrique; Sabater-Navarro, José María
Resumen : A low-cost, additively manufactured, biocompatible&#xD;
glucose sensor based on the changes in the unloaded quality&#xD;
factor (Qu) with a single split-ring resonator is presented.&#xD;
An exposition of the fundamentals for the use of the Qu as&#xD;
sensing parameter with microwave planar resonant sensors is&#xD;
shown. The convenience of this sensing parameter is analyzed&#xD;
from the theoretical point of view, and practical design and&#xD;
optimization guidelines are inferred with a special focus on&#xD;
the optimization of Qu sensitivity to glucose concentration. For&#xD;
practical demonstration and experimental assessment, a novel&#xD;
inverted microstrip configuration is considered, built upon a customized&#xD;
structure made with a certified biocompatible material&#xD;
thanks to 3-D printing techniques, which is aimed to provide&#xD;
for a stronger interaction between the electromagnetic fields and&#xD;
the sample. Two metallization solutions are investigated, yielding&#xD;
devices operating at 4.50 and 4.62 GHz, with operating Qu of&#xD;
16.36 and 22.00, relative Qu sensitivities of 1.377 and 2.727, Qu&#xD;
sensitivities to glucose content within the physiological range of&#xD;
0.3 × 10−3 per mg/dL and 0.6 × 10−3 per mg/dL, sensing areas&#xD;
of approximately 11.7 × 8.8 mm2 and total structure sizes of&#xD;
950.0 × 35.0 × 3.5 mm3. The devices show good performance with&#xD;
water–glucose solutions covering a wide range of concentrations,&#xD;
involving physiological as well as industry-related ones.
Notas: Los derechos de autor de este artículo pertenecen al IEEE. El artículo se encuentra publicado en: https://ieeexplore.ieee.org/abstract/document/9585096</description>
      <pubDate>Mon, 12 Jan 2026 11:35:37 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/38844</guid>
      <dc:date>2026-01-12T11:35:37Z</dc:date>
    </item>
    <item>
      <title>Neural Tract Avoidance Path-Planning Optimization: Robotic Neurosurgery</title>
      <link>https://hdl.handle.net/11000/38843</link>
      <description>Título : Neural Tract Avoidance Path-Planning Optimization: Robotic Neurosurgery
Autor : Manrique-Cordoba, Juliana; Martorell, Carlos; Romero-Ante, Juan D.; Sabater-Navarro, José María
Resumen : Background: We propose a three-dimensional path-planning method to generate optimized&#xD;
surgical trajectories for steering flexible needles along curved paths while avoiding critical tracts in&#xD;
the context of surgical glioma resection. Methods: Our approach is based on an application of the&#xD;
rapidly exploring random tree algorithm for multi-trajectory generation and optimization, with a cost&#xD;
function that evaluates different entry points and uses the information of MRI images as segmented&#xD;
binary maps to compute a safety trajectory. As a novelty, an avoidance module of the critical neuronal&#xD;
tracts defined by the neurosurgeon is included in the optimization process. The proposed strategy&#xD;
was simulated in real-case 3D environments to reach a glioma and bypass the tracts of the forceps&#xD;
minor from the corpus callosum. Results: A formalism is presented that allows for the evaluation of&#xD;
different entry points and trajectories and the avoidance of selected critical tracts for the definition of&#xD;
new neurosurgical approaches. This methodology can be used for different clinical cases, allowing the&#xD;
constraints to be extended to the trajectory generator. We present a clinical case of glioma at the base&#xD;
of the skull and access it from the upper area while avoiding the minor forceps tracts. Conclusions:&#xD;
This path-planning method offers alternative curved paths with which to reach targets using flexible&#xD;
tools. The method potentially leads to safer paths, as it permits the definition of groups of critical&#xD;
tracts to be avoided and the use of segmented binary maps from the MRI images to generate new&#xD;
surgical approaches.</description>
      <pubDate>Mon, 12 Jan 2026 11:29:50 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/38843</guid>
      <dc:date>2026-01-12T11:29:50Z</dc:date>
    </item>
    <item>
      <title>N-Dimensional Reduction Algorithm for Learning from Demonstration Path Planning</title>
      <link>https://hdl.handle.net/11000/38842</link>
      <description>Título : N-Dimensional Reduction Algorithm for Learning from Demonstration Path Planning
Autor : Manrique-Cordoba, Juliana; Casa-Lillo, Miguel Ángel de la; Sabater-Navarro, José María
Resumen : This paper presents an n-dimensional reduction algorithm for Learning from&#xD;
Demonstration (LfD) for robotic path planning, addressing the complexity of highdimensional&#xD;
data. The method extends the Douglas–Peucker algorithm by incorporating&#xD;
velocity and orientation alongside position, enabling more precise trajectory simplification.&#xD;
A magnitude-based normalization process preserves proportional relationships across&#xD;
dimensions, and the reduced dataset is used to train Hidden Markov Models (HMMs),&#xD;
where continuous trajectories are discretized into identifier sequences. The algorithm is&#xD;
evaluated in 2D and 3D environments with datasets combining position and velocity. The&#xD;
results show that incorporating additional dimensions significantly enhances trajectory&#xD;
simplification while preserving key information. Additionally, the study highlights the importance&#xD;
of selecting appropriate encoding parameters to achieve optimal resolution. The&#xD;
HMM-based models generated new trajectories that retained the patterns of the original&#xD;
demonstrations, demonstrating the algorithm’s capacity to generalize learned behaviors&#xD;
for trajectory learning in high-dimensional spaces.</description>
      <pubDate>Mon, 12 Jan 2026 11:28:13 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/38842</guid>
      <dc:date>2026-01-12T11:28:13Z</dc:date>
    </item>
    <item>
      <title>Methods for the Segmentation of Reticular Structures Using çD LiDAR Data: A&#xD;
Comparative Evaluation</title>
      <link>https://hdl.handle.net/11000/36844</link>
      <description>Título : Methods for the Segmentation of Reticular Structures Using çD LiDAR Data: A&#xD;
Comparative Evaluation
Autor : Soler Mora, Francisco J.; Peidró Vidal, Adrián; Fabregat-Jaén, Marcos; Payá Castelló, Luis; Reinoso García, Oscar
Resumen : Reticular structures are the basis of major infrastructure projects, including bridges, electrical pylons&#xD;
and airports. However, inspecting and maintaining these structures is both expensive and hazardous, traditionally&#xD;
requiring human involvement.While some research has been conducted in this  eld of study,most e orts focus on faults&#xD;
identi cation through images or the design of robotic platforms, o en neglecting the autonomous navigation of robots&#xD;
through the structure.  is study addresses this limitation by proposing methods to detect navigable surfaces in truss&#xD;
structures, thereby enhancing the autonomous capabilities of climbing robots to navigate through these environments.&#xD;
 e paper proposesmultiple approaches for the binary segmentation between navigable surfaces and background from&#xD;
çD point clouds captured frommetallic trusses. Approaches can be classi ed into two paradigms: analytical algorithms&#xD;
and deep learning methods. Within the analytical approach, an ad hoc algorithm is developed for segmenting the&#xD;
structures, leveraging di erent techniques to evaluate the eigendecomposition of planar patches within the point cloud.&#xD;
In parallel, widely used and advanced deep learning models, including PointNet, PointNet++, MinkUNetç¥C, and&#xD;
PointTransformerVç, are trained and evaluated for the same task. A comparative analysis of these paradigms reveals&#xD;
some key insights.  e analytical algorithm demonstrates easier parameter adjustment and comparable performance&#xD;
to that of the deep learning models, despite the latter’s higher computational demands. Nevertheless, the deep learning&#xD;
models stand out in segmentation accuracy, with PointTransformerVç achieving impressive results, such as a Mean&#xD;
Intersection Over Union (mIoU) of approximately ÀÞ%.  is study highlights the potential of analytical and deep&#xD;
learning approaches to improve the autonomous navigation of climbing robots in complex truss structures. e  ndings&#xD;
underscore the trade-o s between computational e ciency and segmentation performance, o ering valuable insights&#xD;
for future research and practical applications in autonomous infrastructure maintenance and inspection.</description>
      <pubDate>Fri, 11 Jul 2025 11:59:17 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36844</guid>
      <dc:date>2025-07-11T11:59:17Z</dc:date>
    </item>
    <item>
      <title>Static Early Fusion Techniques for Visible and Thermal Images&#xD;
to Enhance Convolutional Neural Network Detection:&#xD;
A Performance Analysis</title>
      <link>https://hdl.handle.net/11000/36843</link>
      <description>Título : Static Early Fusion Techniques for Visible and Thermal Images&#xD;
to Enhance Convolutional Neural Network Detection:&#xD;
A Performance Analysis
Autor : Heredia-Aguado, Enrique; Cabrera, Juan José; Jiménez, Luis Miguel; Valiente, David; Gil, Arturo
Resumen : This paper presents a comparison of different image fusion methods for matching&#xD;
visible-spectrum images with thermal-spectrum (far-infrared) images, aimed at enhancing&#xD;
person detection using convolutional neural networks (CNNs). While object detection&#xD;
with RGB images is a well-developed area, it is still greatly limited by lighting conditions.&#xD;
This limitation poses a significant challenge in image detection playing a larger role in&#xD;
everyday technology, where illumination cannot always be controlled. Far-infrared images&#xD;
(which are partially invariant to lighting conditions) can serve as a valuable complement to&#xD;
RGB images in environments where illumination cannot be controlled and robust object&#xD;
detection is needed. In this work, various early and middle fusion techniques are presented&#xD;
and compared using different multispectral datasets, with the aim of addressing these&#xD;
limitations and improving detection performance.</description>
      <pubDate>Fri, 11 Jul 2025 11:57:12 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36843</guid>
      <dc:date>2025-07-11T11:57:12Z</dc:date>
    </item>
    <item>
      <title>Detection of UAVs on a collision course using optical flow</title>
      <link>https://hdl.handle.net/11000/36842</link>
      <description>Título : Detection of UAVs on a collision course using optical flow
Autor : Cabrera, Juan José; Gil, Arturo; Payá, Luis; Santo, Antonio; Reinoso, Oscar; Rodríguez, David
Resumen : This paper presents a method to detect, track and predict a potential collision with UAVs using an aircraft equipped with a single camera. The method analyses the movement in the camera’s image plane by means of sparse optical flow. In this way, the camera’s own movement can be modelled and cancelled by estimating a homography matrix from a set of corresponding points. Once the movement caused by the camera is cancelled other moving objects can be isolated and the presence of other UAVs can be detected. Additionally, the method predicts potential collisions by examining the alignment between the position and velocity vectors of the UAV, which are estimated up to a scale factor. The proposed method is effective at detecting and predicting collisions with UAVs, regardless of their appearance, size, or movement, making it useful for applications related to airspace security.</description>
      <pubDate>Fri, 11 Jul 2025 11:56:26 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36842</guid>
      <dc:date>2025-07-11T11:56:26Z</dc:date>
    </item>
    <item>
      <title>Triplet Neural Networks for the Visual Localization of Mobile Robots</title>
      <link>https://hdl.handle.net/11000/36841</link>
      <description>Título : Triplet Neural Networks for the Visual Localization of Mobile Robots
Autor : Alfaro, Marcos; Cabrera, Juan José; Jiménez, Luis Miguel; Reinoso, Óscar; Payá, Luis
Resumen : Triplet networks are composed of three identical convolutional neural networks that function in parallel and&#xD;
share their weights. These architectures receive three inputs simultaneously and provide three different outputs,&#xD;
and have demonstrated to have a great potential to tackle visual localization. Therefore, this paper&#xD;
presents an exhaustive study of the main factors that influence the training of a triplet network, which are the&#xD;
choice of the triplet loss function, the selection of samples to include in the training triplets and the batch size.&#xD;
To do that, we have adapted and retrained a network with omnidirectional images, which have been captured&#xD;
in an indoor environment with a catadioptric camera and have been converted into a panoramic format. The&#xD;
experiments conducted demonstrate that triplet networks improve substantially the performance in the visual&#xD;
localization task. However, the right choice of the studied factors is of great importance to fully exploit the&#xD;
potential of such architectures</description>
      <pubDate>Fri, 11 Jul 2025 11:54:16 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36841</guid>
      <dc:date>2025-07-11T11:54:16Z</dc:date>
    </item>
    <item>
      <title>Evaluation of Open-Source OCR Libraries for Scene Text Recognition in&#xD;
the Presence of Fisheye Distortion</title>
      <link>https://hdl.handle.net/11000/36840</link>
      <description>Título : Evaluation of Open-Source OCR Libraries for Scene Text Recognition in&#xD;
the Presence of Fisheye Distortion
Autor : Flores, María; Valiente, David; Alfaro, Marcos; Fabregat-Jaénn, Marc; Payá, Luis
Resumen : Due to the rich and precise semantic information that text provides, scene text recognition is relevant in a wide&#xD;
range of vision-based applications. In recent years, the use of vision systems that combine a camera and a&#xD;
fisheye lens is common in a variety of applications. The addition of a fisheye lens has the great advantage of&#xD;
capturing a wider field of view, but this causes a great deal of distortion, making certain tasks challenging.&#xD;
In many applications, such as localization or mapping for a mobile robot, the algorithms work directly with&#xD;
fisheye images (i.e. distortion is not corrected). For this reason, the principal objective of this work is to study&#xD;
the effectiveness of some OCR (Optical Character Recognition) open-source libraries applied to images with&#xD;
fisheye distortion. Since no scene text dataset of this kind of image has been found, this work also generates&#xD;
a synthetic image dataset. A fisheye model which varies some parameters is applied to standard images of a&#xD;
benchmark scene text dataset to generate the proposed dataset.</description>
      <pubDate>Fri, 11 Jul 2025 11:52:10 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36840</guid>
      <dc:date>2025-07-11T11:52:10Z</dc:date>
    </item>
    <item>
      <title>Augmented Feasibility Maps: A Simultaneous Approach to Redundancy&#xD;
Resolution and Path Planning</title>
      <link>https://hdl.handle.net/11000/36839</link>
      <description>Título : Augmented Feasibility Maps: A Simultaneous Approach to Redundancy&#xD;
Resolution and Path Planning
Autor : Fabregat-Jaén, Marc; Peidró, Adrián; González-Amorós, Esther; Flores, María; Reinoso, Oscar
Resumen : Redundant robotic manipulators are capable of performing complex tasks with an unprecedented level of&#xD;
dexterity and precision. However, their redundancy also introduces significant computational challenges, particularly&#xD;
in the realms of redundancy resolution and path planning. This paper introduces a novel approach to&#xD;
simultaneously address these challenges through the concept of Augmented Feasibility Maps, by integrating&#xD;
task coordinates as decision variables into the traditional feasibility maps. We validate the AFM concept by using&#xD;
Rapidly-Exploring Random Trees to explore the maps, demonstrating its efficacy in simulations of various&#xD;
dimensionalities. The method is capable of incorporating kinematic constraints, such as obstacle avoidance&#xD;
while adhering to joint limits.</description>
      <pubDate>Fri, 11 Jul 2025 10:08:51 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36839</guid>
      <dc:date>2025-07-11T10:08:51Z</dc:date>
    </item>
    <item>
      <title>A Method for the Calibration of a LiDAR and Fisheye Camera System</title>
      <link>https://hdl.handle.net/11000/36837</link>
      <description>Título : A Method for the Calibration of a LiDAR and Fisheye Camera System
Autor : Martínez Ballester, Álvaro; Santo, Antonio; Ballesta, Monica; Gil, Arturo; Payá, Luis
Resumen : LiDAR and camera systems are frequently used together to gain a more complete&#xD;
understanding of the environment in different fields, such as mobile robotics, autonomous&#xD;
driving, or intelligent surveillance. Accurately calibrating the extrinsic parameters is crucial&#xD;
for the accurate fusion of the data captured by both systems, which is equivalent to finding&#xD;
the transformation between the reference systems of both sensors. Traditional calibration&#xD;
methods for LiDAR and camera systems are developed for pinhole cameras and are not&#xD;
directly applicable to fisheye cameras. This work proposes a target-based calibration&#xD;
method for LiDAR and fisheye camera systems that avoids the need to transform images&#xD;
to a pinhole camera model, reducing the computation time. Instead, the method uses the&#xD;
spherical projection of the image, obtained with the intrinsic calibration parameters and the&#xD;
corresponding point cloud for LiDAR–fisheye calibration. Thus, unlike a pinhole-camerabased system, a wider field of view is provided, adding more information, which will lead&#xD;
to a better understanding of the environment itself, as well as enabling using fewer image&#xD;
sensors to cover a wider area.</description>
      <pubDate>Fri, 11 Jul 2025 09:44:36 GMT</pubDate>
      <guid isPermaLink="false">https://hdl.handle.net/11000/36837</guid>
      <dc:date>2025-07-11T09:44:36Z</dc:date>
    </item>
  </channel>
</rss>

