Publications

publications by categories in reversed chronological order. generated by jekyll-scholar.

2025

  1. NeurIPS
    RtRank: Stratified Response Time Ranking for Data-Efficient Reward Modeling
    Timo KaufmannYannick MetzDaniel Keim, and 1 more author
    In NeurIPS accepted, to be published., 2025
  2. TMLR
    A Survey of Reinforcement Learning from Human Feedback
    Timo KaufmannPaul WengViktor Bengs, and 1 more author
    Transactions on Machine Learning Research, 2025
  3. arXiv
    Feedback Forensics: A Toolkit to Measure AI Personality
    Arduin FindeisTimo KaufmannEyke Hüllermeier, and 1 more author
    2025
  4. ICLR
    Inverse Constitutional AI: Compressing Preferences into Principles
    Arduin FindeisTimo KaufmannEyke Hüllermeier, and 2 more authors
    In Proceedings of the International Conference on Learning Representations (ICLR), 2025
  5. ICML
    Comparing Comparisons: Informative and Easy Human Feedback with Distinguishability Queries
    Xuening Feng, Zhaohui Jiang, Timo Kaufmann, and 3 more authors
    In Proceedings of the International Conference on Machine Learning (ICML), 2025
  6. AAAI
    DUO: Diverse, Uncertain, On-Policy Query Generation and Selection for Reinforcement Learning from Human Feedback
    Xuening Feng, Zhaohui Jiang, Timo Kaufmann, and 4 more authors
    In Proceedings of the AAAI Conference on Artificial Intelligence, 2025
  7. CL
    Problem Solving Through Human-AI Preference-Based Cooperation
    Subhabrata DuttaTimo Kaufmann, Goran Glavaš, and 7 more authors
    Computational Linguistics, 2025

2024

  1. MHFAIA
    Comparing Comparisons: Informative and Easy Human Feedback with Distinguishability Queries
    Xuening Feng, Zhaohui Jiang, Timo Kaufmann, and 3 more authors
    In ICML 2024 Workshop on Models of Human Feedback for AI Alignment (MHFAIA), 2024
  2. MHFAIA
    Relatively Rational: Learning Utilities and Rationalities Jointly from Pairwise Preferences
    Taku YamagataTobias OberkoflerTimo Kaufmann, and 3 more authors
    In ICML 2024 Workshop on Models of Human Feedback for AI Alignment (MHFAIA), 2024
  3. RLBRew
    OCALM: Object-Centric Assessment with Language Models
    Timo KaufmannJannis BlümlAntonia Wüst, and 3 more authors
    In RLC 2024 Workshop on Reinforcement Learning Beyond Rewards (RLBRew), 2024

2023

  1. HLDM
    On the Challenges and Practices of Reinforcement Learning from Real Human Feedback
    Timo KaufmannSarah BallJacob Beck, and 2 more authors
    In Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2023
  2. Reinforcement Learning from Human Feedback for Cyber-Physical Systems: On the Potential of Self-Supervised Pretraining
    Timo KaufmannViktor Bengs, and Eyke Hüllermeier
    In Proceedings of the International Conference on Machine Learning for Cyber-Physical Systems (ML4CPS), 2023