We’ve finally put all the videos, slides and other resources together from the Best Practices in AI Afternoon event that happened on the 5th of July 2024. You can find them listed below.
Karin Sevegnani, Senior Solutions Architect, Nvidia
In this talk, we will explore the intricate balance between computational resources, memory limitations, and parameter-efficient fine-tuning techniques in large language models (LLMs). We will analyse strategies to optimize the performance of LLMs while managing these constraints effectively. From efficient memory utilization to streamlined parameter fine-tuning methods, we will discuss practical approaches to maximize the efficiency of LLMs without sacrificing performance.
Ryan Daniels, University of Cambridge
Writing research software in Python presents numerous challenges to reproducibility - what version of Python is being used? What about the versions of PyTorch, Scikit Learn or Numpy? Should we use Conda, or venv, or Poetry to manage dependencies and environments? How can we control randomness? Do I have the right version of Cuda Toolkit? In principle, given the same data, and same algorithms and methodology, we should be able to reproduce the results of any given experiment to within an acceptable degree of error. Dealing with the above questions introduces significant problems to reproducing experiments in machine learning. In this talk, I would like to convince you that Docker can help alleviate almost all of these questions. Furthermore, combining Docker, git and GitHub can be a powerful workflow, helping to minimise your tech stack, and declutter your python development experience.
Wahab Kawafi, University of Bristol
Covering methods such as mock testing, simulation, experiment tracking, and dataset curation. With examples in medicine, chemistry, aerospace engineering, and LLMs.
Edwin Brown, Research Software Engineering, University of Sheffield
Practical guide to profile machine learning code to find bottlenecks and to remove these bottlenecks.
Denis Battistella, Higher Education and Research, Nvidia
Links to the resources discussed in the presentation:
We’d like to say thank you to all the speakers and attendees for making Best Practices in AI Afternoon a great success!
Thank you to Emma and Kate from the Centre for Machine Intelligence (CMI) for all the help with organising the event and the CMI and Nvidia for sponsoring the event.
Karin Sevegnani presenting “Maximizing Efficiency in Large Language Models: Compute, Memory, and Fine-Tuning”
Ryan Daniels presenting “Docker for Machine Learning”
Edwin Brown presenting “How to make your machine learning code faster”
Bob Turner presenting “From Research Software to Software as a Service”
Denis Battistella presenting “Nvidia Self-paced Training Courses”
Q & A Panel with Andy Grant, Edwin Brown, Ryan Daniels, and Christopher Wild
For queries relating to collaborating with the RSE team on projects: rse@sheffield.ac.uk
Information and access to JADE II and Bede.
Join our mailing list so as to be notified when we advertise talks and workshops by subscribing to this Google Group.
Queries regarding free research computing support/guidance should be raised via our Code clinic or directed to the University IT helpdesk.