Post-doc position on Large Language Models for Low-Resource Languages
This position is part of an ambitious initiative of the HiTZ center, funded, among others, by the Basque and Spanish Governments, to improve the performance of Large Language Models on languages with low-resource corpora, including Basque. The motivation is that the amount of digital text, instructions and preference data available for most of the languages is orders of magnitude smaller than the text available for the largest languages. Current techniques for building Large Language Models like GPT or Llama require massive amounts of text, and if fact, their performance for smaller languages is significantly worse.
The research topic is relatively open as far as it looks for sinergies with the aforementioned projects.
The researcher will join a thriving team of other PhDs and researchers focused on the limitations of current Large Language Models like GPT, and how to overcome them. As a sample please check the recently accepted ACL paper at arxiv, as well as others in our website.
The candidate should preferably have a BSc degree in computer science, telecommunications engineering, mathematics or physics, and a PhD in language technologies and/or machine learning. We are looking for individuals who are passionate about natural language processing and have a strong background in computer science and related fields. The applicants must demonstrate excellent communication skills in English.
Our ideal candidate has experience in machine learning, deep learning, and statistical analysis, as well as a strong proficiency in programming languages such as Python.
We welcome applicants from all backgrounds and are committed to creating an inclusive and supportive workplace.
2 years
Aproximately 40,640 euros per year.
The advisors will be Mikel Artetxe and Eneko Agirre. If you have any question, please do not hesitate to contact us at this address: recruitment.hitz@ehu.eus. Please include the job ID when contacting us.
To submit your application please follow this link.