
Parisa Safikhani
Research Area Research Infrastructure and Methods
Researcher
- +49 511 450670-468
List of projects
List of publications
7 Übereinstimmungen gefunden /
Context-aware search space adaptation of hyperparameters and architectures for AutoML in text classification.Safikhani, P., & Broneske, D. (2025).Context-aware search space adaptation of hyperparameters and architectures for AutoML in text classification. ACL Anthology, 1018-1027. Abstract
While Automated Machine Learning (AutoML) systems have shown strong performance on structured data, their application to natural language processing (NLP) tasks remains limited by static, task-agnostic search spaces. In this work, we propose a context-aware extension of AutoPyTorch that dynamically adapts both the hyperparameter search space and neural architecture configuration based on corpus-level meta-features. Our approach extracts interpretable textual statistics—such as average sequence length, vocabulary richness, and class imbalance—to guide the configuration of key hyperparameters. We also introduce two adaptive neural backbones, whose structures are shaped by these meta-features to improve model expressiveness and generalization. |
Static and dynamic contextual embedding for AutoML in text classification tasks.Safikhani, P., & Broneske, D. (2025).Static and dynamic contextual embedding for AutoML in text classification tasks. In IEEE Institute of Electrical and Electronic Engineers (Hrsg.), 2025 7th International Conference on Natural Language Processing (ICNLP) (S. 292-301). Jacksonville, Florida, USA: IEEE Xplore. https://doi.org/10.1109/ICNLP65360.2025.11108687 |
AutoML meets hugging face: Domain-aware pretrained model selection for text classification.Safikhani, P., & Broneske, D. (2025).AutoML meets hugging face: Domain-aware pretrained model selection for text classification. In A. Ebrahimi, S. Haider, E. Liu, M. L. Pacheco, & S. Wein (Hrsg.), Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 4: Student Research Workshop). Albuquerque, USA: Association for Computational Linguistics. Abstract
The effectiveness of embedding methods is crucial for optimizing text classification performance in Automated Machine Learning (AutoML). However, selecting the most suitable pre-trained model for a given task remains challenging. This study introduces the Corpus-Driven Domain Mapping (CDDM) pipeline, which utilizes a domain-annotated corpus of pre-fine-tuned models from the Hugging Face Model Hub to improve model selection. Integrating these models into AutoML systems significantly boosts classification performance across multiple datasets compared to baseline methods. Despite some domain recognition inaccuracies, results demonstrate CDDM’s potential to enhance model selection, streamline AutoML workflows, and reduce computational costs. |
Framing and BERTology: A data-centric approach to integration of linguistic features into transformer-based pre-trained language models.Avetisyan, H., Safikhani, P., & Broneske, D. (2024).Framing and BERTology: A data-centric approach to integration of linguistic features into transformer-based pre-trained language models. In Arai, K. (Hrsg.), Intelligent Systems and Applications (S. 81-90). Cham: Springer. https://doi.org/10.1007/978-3-031-47718-8 |
Laughing out loud – Exploring AI-generated and human-generated humor.Safikhani, P., Avetisyan, H., & Broneske, D. (2023).Laughing out loud – Exploring AI-generated and human-generated humor. Computer Science & Information Technology (CS & IT), 2023, 59-76. https://doi.org/10.5121/csit.2023.132406 |
Enhancing AutoNLP with fine-tuned BERT models: An evaluation of text representation methods for AutoPyTorch.Safikhani, P., & Broneske, D. (2023).Enhancing AutoNLP with fine-tuned BERT models: An evaluation of text representation methods for AutoPyTorch. In D. C. Wyld & D. Nagamalai (Hrsg.), Computer Science & Information Technology (CS & IT) (S. 23-38). Chennai, Tamil Nadu, India: AIRCC Publishing Corporation. |
Automated occupation coding with hierarchical features: A data-centric approach to classification with pre-trained language models.Safikhani, P., Avetisyan, H., Föste-Eggers, D., & Broneske, D. (2023).Automated occupation coding with hierarchical features: A data-centric approach to classification with pre-trained language models. Discover Artificial Intelligence 3, 2023(6). https://doi.org/10.1007/s44163-023-00050-y |
List of presentations & conferences
10 Übereinstimmungen gefunden /