Implementing fast large language models (LLM) in Codon programming language (Implementacija velikih jezičkih modela u Codon programskom jeziku)

Period trajanja: 1. 6. 2024. 31. 12. 2024.
Rok za prijave: 23. 4. 2024. 10. 5. 2024.

Python is arguably the most popular programming language nowadays, and is thus widely used for machine learning research. Recently, the hottest topic in machine learning are large language models (LLMs) that are used for general-purpose language generation and various other natural language processing tasks such as machine translation. However, the base Python is not an ideal language for implementing those models due to its performance; thus, the most performant LLMs are currently implemented in C, C++ or Rust. We have recently developed Codon, a static ahead-of-time compiler for Python, and a highly optimized implementation of NumPy3 that enables Python applications to achieve the speed of C implementations with ease. In this project, we are looking to apply Codon and its NumPy libraries to implement various LLMs, such as Gemini, Llama, Mixtral, and to compare the new implementations with the current state-of-the-art implementations in other languages. 


● Read and understand the literature

● Explore various LLMs (Gemini, Llama, Mixtral etc.)

● Implement these LLMs in Codon

● Compare the accuracy and the performance between the reference LLM implementations and the Codon implementations

● Write a report detailing the improvements and the differences

Ibrahim Numanagić

Pretplatite se na novosti o BH Akademskom Imeniku

Ova stranica koristi kolačiće da bi vam pružila najbolje iskustvo

Saznaj više