Consent

This site uses third party services that need your consent.

Jahid Hasan Polash (B.Sc.)

Research Assistant
ORCID: 0009-0000-0099-9238

Biography

Jahid Hasan Polash is currently completing his Master of Science in Web and Data Science at the University of Koblenz, where his thesis work examines information gaps in biomedical case studies through analysis of patient-reported Reddit posts. Prior to his current position, he served as a Research Assistant at the Fraunhofer Institute for Applied Information Technology from April 2024 to October 2025.

Jahid holds a Bachelor of Science in Computer Science and Engineering from Shahjalal University of Science and Technology in Bangladesh, where he graduated in 2017. His technical background encompasses database systems, machine learning frameworks, and biomedical text analysis, positioning him at the forefront of applying advanced computational methods to healthcare challenges.

Contact

Academic Background

Areas of Expertise

Data Scientist || Software Engineer

Publications from Jahid Hasan Polash

ELMTEX: Fine-Tuning LLMs for Structured Clinical Information Extraction. A Case Study on Clinical Reports

Aynur Guluzade, Naguib Heiba, Zeyd Boukhers, Florim Hamiti, Jahid Hasan Polash, Yehya Mohamad & Carlos A. Velasco

Europe’s healthcare must improve interoperability and embrace solutions to unlock the value of legacy clinical data. We used LLMs to transform unstructured clinical reports into structured records. We built a complete workflow, including a UI, and benchmarked various LLM sizes through both prompt engineering and fine-tuning. Our fine-tuned smaller models matched or even surpassed the larger ones, making them ideal for settings with limited computational resources. Finally, we validated a novel dataset of annotated English and German translations of clinical summaries using automated metrics alongside expert manual review.

Reinforcement Learning for Large Language Model Fine-Tuning: A Systematic Literature Review

Lingxiao Kong, Qusai Ramdan, Oussama Zoubia, Jahid Hasan Polash, Mayra Elwes, Mehdi Akbari Gurabi, Lu Jin, Ekaterina Kutafina, Roman Matzutt, Yuanbin Wang, Junqi Xu, Oya Deniz Beyan, Cong Yang, Zeyd Boukhers

Large Language Models (LLMs) have been developed for a wide range of language-based tasks, while Reinforcement Learning (RL) has been primarily applied to decision-making problems such as robotics, game theory, and control systems. Nowadays, these two paradigms are integrated through different synergies. In this literature review, we focus on \textit{RL4LLM fine-tuning}, where RL techniques are systematically leveraged to fine-tune LLMs and align them with various preferences. Our review provides a comprehensive analysis of 230 recent publications, presenting a methodological taxonomy that organizes current research into three primary method domains: \textit{Optimization Algorithm}, concerning innovation in core RL update rules; \textit{Training Framework}, regarding innovation in the orchestration of the training process; and \textit{Reward Modeling}, addressing how LLMs learn and represent preferences and feedback. Within these primary domains, we further analyze methods and innovations through more granular categories to provide an in-depth summary of RL4LLM fine-tuning research. We address three research questions: 1) recent methods overview, 2) methodological innovations, and 3) limitations and future directions. Our analysis comprehensively demonstrates the breadth and impact of recent RL4LLM fine-tuning research while highlighting valuable directions for future investigation.