Benchmarking ethical reasoning in Large Language Models (Project #6)

University of Oslo, Department of Informatics 

Three year PhD position
(An extension of the appointment by up to twelve additional months may be considered, which will be devoted to career-enhancing work duties, e.g. teaching or supervision. This will be dependent on the qualifications of the applicant and the specific teaching needs of the department.)

Description

This project will take a cross-disciplinary approach to the design and use of benchmarks for evaluating ethical and moral reasoning in Large Language Models (LLMs). Combining approaches from experimental philosophy and natural language processing, the project will address research questions like the following:  

  • How can we best design benchmarks for evaluating ethical reasoning in LLMs? What are the strengths and weaknesses of current approaches?  
  • How do the different stages in the LLM training pipeline or different sources of training data impact the knowledge learned about ethics?  
  • How robust and consistent are the moral judgments of a given LLM? This could be investigated with respect to, e.g., word choice or framing, and cross-lingual or cross-cultural effects.  

This project forms part of a collaboration between researchers from several departments at the University of Oslo, including the Language Technology Group at the Department of Informatics, the Practical Philosophy Group at the Department of Philosophy, Classics, History of Art and Ideas, and the Department of Biostatistics at the Institute of Basic Medical Sciences.

Specific project requirements

  • Master’s degree in computer science with a specialization in Natural Language Processing, including background from neural Machine Learning.

  • Demonstrable experience of/interest in philosophy and ethics. A formal qualification in philosophy is highly advantageous, but not a requirement.

Supervisors

 

Published Jan. 29, 2024 9:34 PM - Last modified Jan. 29, 2024 9:34 PM