Sorry, you need to enable JavaScript to visit this website.

Deepfake detection is critical in mitigating the societal threats posed by manipulated videos. While various algorithms have been developed for this purpose, challenges arise when detectors operate externally, such as on smartphones, when users take a photo of deepfake images and upload on the Internet. One significant challenge in such scenarios is the presence of Moire patterns, which degrade image quality and confound conventional classification algorithms, including deep neural networks (DNNs). The impact of Moire patterns remains largely unexplored for deepfake detectors.

Categories:
16 Views

The emerging deep learning algorithms have shown significant potential in the development of efficient computer aided diagnosis tools for automated detection of lung infections using chest radiographs. However, many existing methods are slice-based and require manual selection of appropriate slices from the entire CT scan, which is tedious and requires expert radiologists.

Categories:
11 Views

Recent advancements in artificial intelligence algorithms for medical imaging show significant potential in automating the detection of lung infections from chest radiograph scans. However, current approaches often focus solely on either 2-D or 3-D scans, failing to leverage the combined advantages of both modalities. Moreover, conventional slice-based methods place a manual burden on radiologists for slice selection.

Categories:
16 Views

In this work, we propose a semi-supervised deep feature generation network that accounts for local similarities. It is based on the deep dictionary learning (DDL) framework. The formulation accounts for two unique aspects of hyperspectral classification. First, the fact that the total number of pixels / samples to be labeled is constant; this allows for a semi-supervised formulation allowing only a few pixels / samples to be labeled as training data. Second, the samples / pixels are spatially correlated; this leads to a graph regularization formulation.

Categories:
9 Views

Although transformer-based models have improved the state-of-the-art in speech recognition, it is still not well understood what information from the speech signal these models encode in their latent representations. This study investigates the potential of using labelled data (TIMIT) to probe wav2vec 2.0 embeddings for insights into the encoding and visualisation of speech signal information at phone boundaries. Our experiment involves training probing models to detect phone-specific articulatory features in the hidden layers based on IPA classifications.

Categories:
46 Views

To alleviate the negative impacts of noisy labels, most of the noisy label learning (NLL) methods dynamically divide the training data into two types, “clean samples” and “noisy samples”, in the training process. However, the conventional selection of clean samples heavily depends on the features learned in the early stages of training, making it difficult to guarantee the cleanliness of the selected samples in scenarios where the noise ratio is high.

Categories:
69 Views

Although BP decoders are efficient and provide significant
performance for classical low-density parity-check (LDPC)
codes, they will suffer a degradation in performance for quantum
LDPC (QLDPC) codes due to the limitations in the quantum
field. In this paper, we propose a posterior adjustment of
either a single qubit or multiple qubits within binary belief
propagation (BP). The adjustment process changes the posterior
likelihood ratio for one or multiple qubits according to the

Categories:
79 Views

Efficient utilization of large-scale biobank data is crucial for inferring the genetic basis of disease and predicting health outcomes from the DNA. Yet we lack efficient, accurate methods that scale to data where electronic health records are linked to whole genome sequence information. To address this issue, our paper develops a new algorithmic paradigm based on Approximate Message Passing (AMP), which is specifically tailored for genomic prediction and association testing.

Categories:
14 Views

Loihi 2 is a fully event-based neuromorphic processor that supports a wide range of synaptic connectivity configurations and temporal neuron dynamics. Loihi 2's temporal and event-based paradigm is naturally well-suited to processing data from an event-based sensor, such as a Dynamic Vision Sensor (DVS) or a Silicon Cochlea. However, this begs the question: How general are signal processing efficiency gains on Loihi 2 versus conventional computer architectures?

Categories:
34 Views

We consider the problem of classifying a pair of Synthetic Aperture Radar (SAR) images by proposing an explainable and frugal algorithm that integrates a set of divergences. The approach relies on a statistical framework that takes standard probability distributions into account for modelling SAR data. Then, by learning a combination of parameterized Renyi divergences and their parameters from the data, we are able to classify the pair of images with fewer parameters than regular machine learning approaches while also allowing an interpretation of the results related to the priors used.

Categories:
21 Views

Pages