publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2023
- BigData 2023Mask-cscGAN for realistic synthetic cell generationPanagiotis Antoniadis, Christina Sartzetaki, Nick Antonopoulos, and 3 more authorsIn Proceedings of the IEEE International Conference on Big Data, 2023
Deep learning methods for RNA sequencing data have exploded in the recent years due to the advent of singlecell RNA sequencing (scRNA-seq), which enables the study of multiple cells per-patient simultaneously. However, in the case of rare cell types, data scarcity continues to exist, posing several challenges, while preventing the exploitation of deep learning models’ full predictive power. Generating realistic synthetic cells to augment the data could allow for more informative subsequent downstream analyses. Herein, we introduce Mask-cscGAN, a conditional generative adversarial network (GAN) that generates realistic synthetic cells with desired characteristics managing also to model genes’ sparsity through learning a mask of zeros. Employed for the augmentation of a glioblastoma multiforme (GBM) malignant cells dataset, Mask-cscGAN generates realistic synthetic cells of desired cancer subtypes. Generating cells of a rare cancer subtype, Mask-cscGAN improves the classification performance of the rare cancer subtype by 12.29%. Mask-cscGAN is the first to generate realistic synthetic cells belonging to specified cancer subtypes, and augmentation with Mask-cscGAN outperforms state-of-the-art methods in rare cancer subtype classification.
@inproceedings{10386596, author = {Antoniadis, Panagiotis and Sartzetaki, Christina and Antonopoulos, Nick and Papageorgiou, Pantelis and Korfiati, Aigli and Pitsikalis, Vassilis}, booktitle = {Proceedings of the IEEE International Conference on Big Data}, title = {Mask-cscGAN for realistic synthetic cell generation}, year = {2023}, pages = {4575-4583}, keywords = {Deep learning;Sequential analysis;Ethics;Costs;RNA;Precision medicine;Predictive models;scRNA-seq;synthetic cells generation;cancer subtype classification;GAN;rare cell types}, doi = {10.1109/BigData59044.2023.10386596}, bibtex_show = true, selected = true }
- SMC 2023Beyond Within-Subject Performance: A Multi-Dataset Study of Fine-Tuning in the EEG DomainChristina Sartzetaki, Panagiotis Antoniadis, Nick Antonopoulos, and 4 more authorsIn Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, 2023
There is a critical demand for BCI systems that can swiftly adapt to a new user and at the same time function with any user. We propose a fine-tuning approach for neural networks that serves a dual purpose; first, to minimize calibration times through requiring considerably less data - up to one-sixth - from the target subject than training from scratch, and second, to alleviate cases of user illiteracy by providing a substantial performance boost of over 11% in absolute accuracy from the features learned from other subjects. Ultimately, our adaptation method surpasses standard within-subject performance by a large margin in all subjects. We present ablation studies across three datasets, in which we demonstrate that fine-tuning outperforms other adaptation methods for BCI systems and that what matters most is the quantity of pre-training subjects, rather than their BCI-ability, achieving over 8% absolute increase in classification accuracy when scaling up the order of magnitude. Finally, we compare our approach to the state-of-the-art in EEG-based motor imagery and find it comparable, if not superior, to methods employing far more complex neural networks, obtaining 82.60% and 85.64% within-subject accuracy in the four-class BCIC IV-2a and binary MMI datasets respectively.
@inproceedings{10394335, author = {Sartzetaki, Christina and Antoniadis, Panagiotis and Antonopoulos, Nick and Gkinis, Ioannis and Krasoulis, Agamemnon and Perdikis, Serafeim and Pitsikalis, Vassilis}, booktitle = {Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics}, title = {Beyond Within-Subject Performance: A Multi-Dataset Study of Fine-Tuning in the EEG Domain}, year = {2023}, pages = {4429-4435}, keywords = {Training;Neural networks;Transfer learning;Self-supervised learning;Electroencephalography;Calibration;Standards;BCI;EEG;Motor Imagery;Domain Adaptation;Fine-tuning;BCI-illiteracy}, doi = {10.1109/SMC53992.2023.10394335}, bibtex_show = true, selected = true }
2022
- MTAPA mechanism for personalized Automatic Speech Recognition for less frequently spoken languages: the Greek casePanagiotis Antoniadis, Emmanouil Tsardoulias, and Andreas SymeonidisMultimedia Tools and Applications, 2022
Automatic Speech Recognition (ASR) has become increasingly popular since it significantly simplifies human-computer interaction, providing a more intuitive way of communication. Building an accurate, general-purpose ASR system is a challenging task that requires a lot of data and computing power. Especially for languages not widely spoken, such as Greek, the lack of adequately large speech datasets leads to the development of ASR systems adapted to a restricted corpus and/or for specific topics. When used in specific domains, these systems can be both accurate and fast, without the need for large datasets and extended training. An interesting application domain of such narrow-scope ASR systems is the development of personalized models that can be used for dictation. In the current work we present three personalization-via-adaptation modules, that can be integrated into any ASR/dictation system and increase its accuracy. The adaptation can be applied both on the language model (based on past text samples of the user) as well as on the acoustic model (using a set of user’s narrations). To provide more precise recommendations, clustering algorithms are applied and topic-specific language models are created. Also, heterogeneous adaptation methods are combined to provide recommendations to the user. Evaluation performed on a self-created database containing 746 corpora included in messaging applications and e-mails from the same user, demonstrates that the proposed approach can achieve better results than the vanilla existing Greek models.
@article{Antoniadis2022, author = {Antoniadis, Panagiotis and Tsardoulias, Emmanouil and Symeonidis, Andreas}, title = {A mechanism for personalized Automatic Speech Recognition for less frequently spoken languages: the Greek case}, journal = {Multimedia Tools and Applications}, year = {2022}, volume = {81}, pages = {40635-40652}, issn = {1573-7721}, doi = {10.1007/s11042-022-12953-6}, bibtex_show = true, url = {https://doi.org/10.1007/s11042-022-12953-6}, }
2021
- FG 2021Exploiting Emotional Dependencies with Graph Convolutional Networks for Facial Expression RecognitionPanagiotis Antoniadis, Panagiotis Paraskevas Filntisis, and Petros MaragosIn Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, 2021
Over the past few years, deep learning methods have shown remarkable results in many face-related tasks including automatic facial expression recognition (FER) in-the-wild. Meanwhile, numerous models describing the human emotional states have been proposed by the psychology community. However, we have no clear evidence as to which representation is more appropriate and the majority of FER systems use either the categorical or the dimensional model of affect. Inspired by recent work in multi-label classification, this paper proposes a novel multi-task learning (MTL) framework that exploits the dependencies between these two models using a Graph Convolutional Network (GCN) to recognize facial expressions in-the-wild. Specifically, a shared feature representation is learned for both discrete and continuous recognition in a MTL setting. Moreover, the facial expression classifiers and the valence-arousal regressors are learned through a GCN that explicitly captures the dependencies between them. To evaluate the performance of our method under real-world conditions we perform extensive experiments on the AffectNet and Aff-Wild2 datasets. The results of our experiments show that our method is capable of improving the performance across different datasets and backbone architectures. Finally, we also surpass the previous state-of-the-art methods on the categorical model of AffectNet.
@inproceedings{9667014, author = {Antoniadis, Panagiotis and Filntisis, Panagiotis Paraskevas and Maragos, Petros}, booktitle = {Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition}, title = {Exploiting Emotional Dependencies with Graph Convolutional Networks for Facial Expression Recognition}, year = {2021}, pages = {1-8}, keywords = {Deep learning;Emotion recognition;Image recognition;Databases;Face recognition;Conferences;Psychology}, doi = {10.1109/FG52635.2021.9667014}, bibtex_show = true, selected = true, }
- ICCV 2021An Audiovisual and Contextual Approach for Categorical and Continuous Emotion Recognition In-the-WildPanagiotis Antoniadis, Ioannis Pikoulis, Panagiotis P. Filntisis, and 1 more authorIn Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 2021
In this work we tackle the task of video-based audio-visual emotion recognition, within the premises of the 2nd Workshop and Competition on Affective Behavior Analysis in-the-wild (ABAW2). Poor illumination conditions, head/body orientation and low image resolution constitute factors that can potentially hinder performance in case of methodologies that solely rely on the extraction and analysis of facial features. In order to alleviate this problem, we leverage both bodily and contextual features, as part of a broader emotion recognition framework. We choose to use a standard CNN-RNN cascade as the backbone of our proposed model for sequence-to-sequence (seq2seq) learning. Apart from learning through the RGB input modality, we construct an aural stream which operates on sequences of extracted mel-spectrograms. Our extensive experiments on the challenging and newly assembled Aff-Wild2 dataset verify the validity of our intuitive multi-stream and multi-modal approach towards emotion recognition in-the-wild. Emphasis is being laid on the the beneficial influence of the human body and scene context, as aspects of the emotion recognition process that have been left relatively unexplored up to this point. All the code was implemented using PyTorch and is publicly available.
@inproceedings{Antoniadis_2021_ICCV, author = {Antoniadis, Panagiotis and Pikoulis, Ioannis and Filntisis, Panagiotis P. and Maragos, Petros}, title = {An Audiovisual and Contextual Approach for Categorical and Continuous Emotion Recognition In-the-Wild}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops}, year = {2021}, pages = {3645-3651}, bibtex_show = true, }