Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
45 result(s) for "Bondi, Luca"
Sort by:
How to Achieve High Spatial Resolution in Organic Optobioelectronic Devices?
Light activated local stimulation and sensing of biological cells hold great promise for minimally invasive bioelectronic interfaces. Organic semiconductors are particularly appealing for these applications due to their optoelectronic properties and biocompatibility. This study examines the material properties necessary to localize the optical excitation and achieve optoelectronic transduction with high spatial resolution. Using photovoltage and photocurrent microscopy, we investigate spatial broadening of local optical excitation in Phthalocyanine/3,4,9,10‐Perylenetetracarboxylic diimide (H2PC/PTCDI) planar heterojunctions. Our measurements reveal that resolution losses are tied to the effective diffusion length of charge carriers at the heterojunction. For the H2PC/PTCDI heterojunction, the diffusion length is determined to be λd = 1.5 ± 0.1 µm, attributed to reduced carrier mobility. Covering the heterojunction with poly(3,4‐ethylenedioxythiophene) polystyrene sulfonate (PEDOT:PSS) improves the charge generation performance but increases the carrier diffusion length to λd = 7.0 ± 0.3 µm due to longer lifetime and higher carrier mobility. These findings elucidate the physical mechanisms underlying transduction and provide design principles for organic semiconductor devices aimed at achieving high efficiency and high spatial resolution for wireless and optically activated bioelectronics. This study explores how the optoelectronic properties of organic semiconductors impact on the resolution of light‐activated bioelectronic interfaces. Using photovoltage and photocurrent microscopy techniques, the authors show that H2PC/PTCDI heterojunctions enable a resolution down to 6.5 µm due to low carrier mobility along the heterojunction. Adding a layer of PEDOT:PSS increases charge separation efficiency, but deteriorates the resolution due to larger photocarrier diffusion.
A framework for the acoustic simulation of passing vehicles using variable length delay lines
The sound produced by vehicles driving on roadways constitutes one of the dominant noise sources in urban areas. The impact of traffic noise on human activities and the related investigation on modeling, assessment, and abatement strategies fueled the research on the simulation of the sound produced by individual passing vehicles. Simulators enable in fact to promote a perceptual assessment of the nature of traffic noise and of the impact of single road agents on the overall soundscape. In this work, we present TrafficSoundSim , an open-source framework for the acoustic simulation of vehicles transiting on a road. We first discuss the generation of the sound signal produced by a vehicle, represented as a combination of road/tire interaction noise and engine noise. We then introduce a propagation model based on the use of variable length delay lines, allowing to simulate acoustic propagation and Doppler effect. The proposed simulator incorporates the effect of air absorption and ground reflection, modeled via complex-valued reflection coefficients dependent on the road surface impedance, as well as a model of the directivity of sound sources representing the passing vehicles. The source signal generation and the propagation stages are decoupled, and all effects are implemented using finite impulse response filters. Moreover, no recorded data is required to run the simulation, making the framework flexible and independent on data availability. Finally, to validate the framework capability to accurately simulate passing vehicles, a comparison between synthetic and recorded pass-by events is presented. The validation shows that sounds generated with the proposed method achieve a good match with recorded events in terms of power spectral density and psychoacoustics metrics as well as a perceptually plausible result.
McKean-Vlasov equations with singular coefficients - a review of recent results
This paper focuses on recent works on McKean-Vlasov stochastic differential equations (SDEs) involving singular coefficients. After recalling the classical framework, we review existing recent literature depending on the type of singularities of the coefficients: on the one hand they satisfy some integrability and measurability conditions only, while on the other hand the drift is allowed to be a generalised function. Different types of dependencies on the law of the unknown and different noises will also be considered. McKean-Vlasov SDEs are closely related to non-linear Fokker-Planck equations that are satisfied by the law (or its density) of the unknown. These connections are often established also in this singular setting and will be reviewed here. Important tools for dealing with singular coefficients are also included in the paper, such as Figalli-Trevisan superposition principle, Zvonkin transformation, Markov marginal uniqueness, and stochastic sewing lemma.
Learning Audio Concepts from Counterfactual Natural Language
Conventional audio classification relied on predefined classes, lacking the ability to learn from free-form text. Recent methods unlock learning joint audio-text embeddings from raw audio-text pairs describing audio in natural language. Despite recent advancements, there is little exploration of systematic methods to train models for recognizing sound events and sources in alternative scenarios, such as distinguishing fireworks from gunshots at outdoor events in similar situations. This study introduces causal reasoning and counterfactual analysis in the audio domain. We use counterfactual instances and include them in our model across different aspects. Our model considers acoustic characteristics and sound source information from human-annotated reference texts. To validate the effectiveness of our model, we conducted pre-training utilizing multiple audio captioning datasets. We then evaluate with several common downstream tasks, demonstrating the merits of the proposed method as one of the first works leveraging counterfactual information in audio domain. Specifically, the top-1 accuracy in open-ended language-based audio retrieval task increased by more than 43%.
Learning to Adapt to Domain Shifts with Few-shot Samples in Anomalous Sound Detection
Anomaly detection has many important applications, such as monitoring industrial equipment. Despite recent advances in anomaly detection with deep-learning methods, it is unclear how existing solutions would perform under out-of-distribution scenarios, e.g., due to shifts in machine load or environmental noise. Grounded in the application of machine health monitoring, we propose a framework that adapts to new conditions with few-shot samples. Building upon prior work, we adopt a classification-based approach for anomaly detection and show its equivalence to mixture density estimation of the normal samples. We incorporate an episodic training procedure to match the few-shot setting during inference. We define multiple auxiliary classification tasks based on meta-information and leverage gradient-based meta-learning to improve generalization to different shifts. We evaluate our proposed method on a recently-released dataset of audio measurements from different machine types. It improved upon two baselines by around 10% and is on par with best-performing model reported on the dataset.
Visuo-Acoustic Hand Pose and Contact Estimation
Accurately estimating hand pose and hand-object contact events is essential for robot data-collection, immersive virtual environments, and biomechanical analysis, yet remains challenging due to visual occlusion, subtle contact cues, limitations in vision-only sensing, and the lack of accessible and flexible tactile sensing. We therefore introduce VibeMesh, a novel wearable system that fuses vision with active acoustic sensing for dense, per-vertex hand contact and pose estimation. VibeMesh integrates a bone-conduction speaker and sparse piezoelectric microphones, distributed on a human hand, emitting structured acoustic signals and capturing their propagation to infer changes induced by contact. To interpret these cross-modal signals, we propose a graph-based attention network that processes synchronized audio spectra and RGB-D-derived hand meshes to predict contact with high spatial resolution. We contribute: (i) a lightweight, non-intrusive visuo-acoustic sensing platform; (ii) a cross-modal graph network for joint pose and contact inference; (iii) a dataset of synchronized RGB-D, acoustic, and ground-truth contact annotations across diverse manipulation scenarios; and (iv) empirical results showing that VibeMesh outperforms vision-only baselines in accuracy and robustness, particularly in occluded or static-contact settings.
An In-Depth Study on Open-Set Camera Model Identification
Camera model identification refers to the problem of linking a picture to the camera model used to shoot it. As this might be an enabling factor in different forensic applications to single out possible suspects (e.g., detecting the author of child abuse or terrorist propaganda material), many accurate camera model attribution methods have been developed in the literature. One of their main drawbacks, however, is the typical closed-set assumption of the problem. This means that an investigated photograph is always assigned to one camera model within a set of known ones present during investigation, i.e., training time, and the fact that the picture can come from a completely unrelated camera model during actual testing is usually ignored. Under realistic conditions, it is not possible to assume that every picture under analysis belongs to one of the available camera models. To deal with this issue, in this paper, we present the first in-depth study on the possibility of solving the camera model identification problem in open-set scenarios. Given a photograph, we aim at detecting whether it comes from one of the known camera models of interest or from an unknown one. We compare different feature extraction algorithms and classifiers specially targeting open-set recognition. We also evaluate possible open-set training protocols that can be applied along with any open-set classifier, observing that a simple of those alternatives obtains best results. Thorough testing on independent datasets shows that it is possible to leverage a recently proposed convolutional neural network as feature extractor paired with a properly trained open-set classifier aiming at solving the open-set camera model attribution problem even to small-scale image patches, improving over state-of-the-art available solutions.
How to Achieve High Spatial Resolution in Organic Optobioelectronic Devices?
Light activated local stimulation and sensing of biological cells offers enormous potential for minimally invasive bioelectronic interfaces. Organic semiconductors are a promising material class to achieve this kind of transduction due to their optoelectronic properties and biocompatibility. Here we investigate which material properties are necessary to keep the optical excitation localized. This is critical to single cell transduction with high spatial resolution. As a model system we use organic photocapacitors for cell stimulation made of the small molecule semiconductors H2Pc and PTCDI. We investigate the spatial broadening of the localized optical excitation with photovoltage microscopy measurements. Our experimental data combined with modelling show that resolution losses due to the broadening of the excitation are directly related to the effective diffusion length of charge carriers generated at the heterojunction. With additional transient photovoltage measurements we find that the H2Pc/PTCDI heterojunction offers a small diffusion length of lambda = 1.5 +/- 0.1 um due to the small mobility of charge carriers along the heterojunction. Instead covering the heterojunction with a layer of PEDOT:PSS improves the photocapacitor performance but increases the carrier diffusion length to lambda = 7.0 +/- 0.3 um due to longer lifetime and higher carrier mobility. Furthermore, we introduce electrochemical photocurrent microscopy experiments to demonstrate micrometric resolution with the pn-junction under realistic aqueous operation conditions. This work offers valuable insights into the physical mechanisms governing the excitation and transduction profile and provide design principles for future organic semiconductor junctions, aiming to achieve high efficiency and high spatial resolution.
Knowledge-driven Scene Priors for Semantic Audio-Visual Embodied Navigation
Generalisation to unseen contexts remains a challenge for embodied navigation agents. In the context of semantic audio-visual navigation (SAVi) tasks, the notion of generalisation should include both generalising to unseen indoor visual scenes as well as generalising to unheard sounding objects. However, previous SAVi task definitions do not include evaluation conditions on truly novel sounding objects, resorting instead to evaluating agents on unheard sound clips of known objects; meanwhile, previous SAVi methods do not include explicit mechanisms for incorporating domain knowledge about object and region semantics. These weaknesses limit the development and assessment of models' abilities to generalise their learned experience. In this work, we introduce the use of knowledge-driven scene priors in the semantic audio-visual embodied navigation task: we combine semantic information from our novel knowledge graph that encodes object-region relations, spatial knowledge from dual Graph Encoder Networks, and background knowledge from a series of pre-training tasks -- all within a reinforcement learning framework for audio-visual navigation. We also define a new audio-visual navigation sub-task, where agents are evaluated on novel sounding objects, as opposed to unheard clips of known objects. We show improvements over strong baselines in generalisation to unseen regions and novel sounding objects, within the Habitat-Matterport3D simulation environment, under the SoundSpaces task.
Training Strategies and Data Augmentations in CNN-based DeepFake Video Detection
The fast and continuous growth in number and quality of deepfake videos calls for the development of reliable detection systems capable of automatically warning users on social media and on the Internet about the potential untruthfulness of such contents. While algorithms, software, and smartphone apps are getting better every day in generating manipulated videos and swapping faces, the accuracy of automated systems for face forgery detection in videos is still quite limited and generally biased toward the dataset used to design and train a specific detection system. In this paper we analyze how different training strategies and data augmentation techniques affect CNN-based deepfake detectors when training and testing on the same dataset or across different datasets.