Michał Tyrolski

Michał Tyrolski

Senior AI Consultant at Ernst & Young / Independent Researcher

Profile My name is Michał Tyrolski. I’m a Senior AI Consultant at Ernst & Young where I work on Agentic AI. I hold an M.Sc. in Machine Learning (2023), supervised by Prof. Piotr Miłoś and Prof. Marek Cygan, and a B.Sc. in Computer Science (2021), supervised by Prof. Henryk Michalewski and Prof. Łukasz Kaiser. Both degrees are from the University of Warsaw, MIM Faculty.

Research Focus My research interests lie at the intersection of AI-based Planning, decision-making in complex environments, and Reinforcement Learning. I am driven by the challenge of advancing the reasoning capabilities of AI algorithms and exploring the strategic thinking that emerges within Game Intelligence. To further pursue this direction, I have recently begun delving in Meta-RL, and Continual Learning to develop algorithms capable of solving problem instances under, broadly speaking, heavy domain shifts across different axes.

Professional Experience

  • Senior AI Consultant IV, Ernst & Young
    2025 - present
    Warsaw, Poland
  • Deep Learning Researcher, DeepFlare
    2022 - 2025
    Warsaw, Poland
  • Data Scientist Intern, Microsoft
    Jul 2022 - Oct 2022
    Dublin, Ireland
  • Teaching Assistant, Uni of Warsaw
    2021 - 2022
    Warsaw, Poland
  • Deep Learning Intern, Nvidia
    Jul 2021 - Oct 2021
    Warsaw, Poland
  • Software Engineering Intern, Microsoft
    Apr 2021 - Jun 2021
    Dublin, Ireland
  • Deep Learning Intern, Nvidia
    Jun 2020 - Sep 2020
    Warsaw, Poland
  • Software Developer Intern AI, Samsung
    Jul 2019 - Sep 2019
    Warsaw, Poland

Education

  • M.Sc. Machine Learning, University of Warsaw
    Warsaw, Poland 2021 - 2023
    top 5% students · graduated with honors
    Key activities:
    • ICLR A* oral top-5%, onsite presentation on RL and planning
    • co-leader ML in PL Conference 2021
  • B.Sc. Computer Science, University of Warsaw
    Warsaw, Poland 2018 - 2021
    top 5% students
    Key activities:
    • Published paper on efficient transformers with Google Research.
    • President, Machine Learning Society at UW.

Computer Science Background My experience spans the entire AI stack: from low-level C++/CUDA engineering (NVIDIA Triton Inference Server) and scalable distributed systems (Microsoft Omex), through MLOps and ML library development (CaRL: Deep RL library calibrated for planning and search), to advanced applied & research innovation across industry and academia. I have worked on deep learning for vaccine discovery (DeepFlare), 3D computer vision algorithms (Samsung), and built my own fastest model parallelism algorithm at the time for extremely large NLP models (NVIDIA). I was the first Microsoft Ireland intern to have a paper accepted at MLADS on explainable AI. My bachelor’s thesis with Google Research set state-of-the-art benchmarks for efficient transformers in long-sequence prediction. During my master’s, I developed Adaptive Subgoal Search (AdaSubS), a novel search algorithm for efficient reinforcement learning under low computational budgets-presented onsite as an ICLR 2023 Top-5% Oral (first such achievement from Poland). Most recently, my research on hierarchical search landscapes was awarded Best Poster at EEML 2025.

Community I am an active member of the AI community, particularly within the ML in PL Association, a non-profit advancing the machine learning community in Poland and across Central & Eastern Europe. Since 2020, I have served as Scientific Program Officer across six annual editions-curating a high-impact invited speaker lineup-and had the honor to be co-Leader of the ML in PL 2021 Conference. In my free time, I enjoy mountain hiking and motorization.

You can view my full CV here.

Selected Publications

OpenGVL: Benchmarking Visual Temporal Progress for Data Curation
CoRL 2025 workshop
Paper Code Website Colab
Benchmark and toolkit for evaluating VLMs' sense of progress in robotics via Value-Order Correlation (VOC); enables automated dataset curation from videos.
Cite
Budzianowski, P., Wiśnios, E., Góral, G., Tyrolski, M., Kulakov, I., Petrenko, V., & Walas, K. (2025). OpenGVL - Benchmarking Visual Temporal Progress for Data Curation. arXiv:2509.17321.
OpenGVL: Benchmarking Visual Temporal Progress for Data Curation
What Matters in Hierarchical Search for Combinatorial Problems?
ICLR 2024 (Generative Models for Decision Making)
Paper Code
Empirical analysis of properties shaping hierarchical search in combinatorial reasoning; guidelines for robust comparison + future design.
Cite
Zawalski, M., Góral, G., Tyrolski, M., Wiśnios, E., Budrowski, F., Cygan, M., Kuciński, Ł., & Miłoś, P. (2024). What matters in hierarchical search for combinatorial reasoning problems? arXiv preprint arXiv:2406.03361.
What Matters in Hierarchical Search for Combinatorial Problems?
Adjusting Planning Horizon with Adaptive Subgoal Search
ICLR 2023 (Top-5%, Oral)
Paper Code Website Colab
AdaSubS adaptively adjusts planning horizon via diverse subgoals + fast reachability filtering; efficient on Sokoban, Rubik's Cube, INT.
Adjusting Planning Horizon with Adaptive Subgoal Search
Hierarchical Transformers Are More Efficient Language Models
NAACL 2022
Paper
Hourglass: a hierarchical Transformer with down/upsampling layers that improves long-sequence modeling efficiency; SOTA on ImageNet32 and strong enwik8 performance.
Cite
Nawrot, P., Tworkowski, S., Tyrolski, M., Kaiser, Ł., Wu, Y., Szegedy, C., & Michalewski, H. (2021). Hierarchical transformers are more efficient language models. arXiv preprint arXiv:2110.13711.
Hierarchical Transformers Are More Efficient Language Models

Other Papers

Explainable Machine Learning at Microsoft MLADS
Microsoft MLADS 2022

Had the privilege to lead explainable AI research at Microsoft Ireland during an internship, utilizing petabytes of data to develop interpretable machine learning solutions for both theoretical and practical applications, presented at the internal MLADS conference.
Cite
Tyrolski, M. (2022). Explainable machine learning at Microsoft MLADS (Internal Microsoft MLADS Conference presentation).
Explainable Machine Learning at Microsoft MLADS
Enhancing Antigenic Peptide Discovery
ICLR 2023, MLDD Workshop
Paper
This study pinpoints evaluation pitfalls in MHC-I presentation prediction and proposes a unified framework to standardize methodology. It also introduces a transformer model trained on interspecies data, markedly improving peptide–MHC-I binding accuracy and generalization across diverse peptides, alleles, and proteins.
Cite
Giziński, S., Preibisch, G., Kucharski, P., Tyrolski, M., Rembalski, M., Grzegorczyk, P., & Gambin, A. (2024). Enhancing antigenic peptide discovery: Improved MHC-I binding prediction and methodology. Methods, 224, 1–9. https://doi.org/10.1016/j.ymeth.2024.01.016
Enhancing Antigenic Peptide Discovery

Selected Projects

OpenGVL: Benchmarking Visual Temporal Progress for Data Curation
Co-author & contributor
OpenGVL

OpenGVL is an open benchmark and toolkit for measuring how well vision–language models understand temporal task progress in robotics, enabling automatic dataset curation by predicting per-frame completion from videos.

  • Problem: Equip robots with a sense of progress for better learning and decision-making.
  • Metric (VOC): Value‑Order Correlation - Spearman rank correlation between predicted progress ordering and true time order.
  • Few‑shot prompting: Uses ordered context episodes to guide predictions on shuffled frames.
  • Contamination control: Hidden tasks with curated demos; 100% completion rate datasets used for evaluation.
  • Unified interface: Standardized prompts, data loaders, and configs (Hydra) across open and closed VLMs.

CaRL Library: Combinatorial RL for planning
Lead author & maintainer
CaRL Architecture

CaRL is an open-source library for scalable offline and online reinforcement/imitation learning in combinatorial planning problems.

  • Supports environments like Sokoban, NPuzzle, Rubik, and INT.
  • Includes 35+ open-source models (Generator, Value, Policy, CLLP).
  • Enables distributed experiments on SLURM clusters and local machines.
  • Interactive Jupyter notebooks for research and reproducibility.
  • Used in multiple peer-reviewed papers.
Key features: modular architecture, Hydra config extension, heterogeneous job support, remote deployment, and dataset demos.