Best Practices in AI Afternoon Event Summary

Twin Karmakharm and Christopher Wild
31 July 2024 10:00
Best Practices in AI Afternoon Banner

We’ve finally put all the videos, slides and other resources together from the Best Practices in AI Afternoon event that happened on the 5th of July 2024. You can find them listed below.

Maximizing Efficiency in Large Language Models: Compute, Memory, and Fine-Tuning

Speaker Karin Sevegnani, Senior Solutions Architect, Nvidia

In this talk, we will explore the intricate balance between computational resources, memory limitations, and parameter-efficient fine-tuning techniques in large language models (LLMs). We will analyse strategies to optimize the performance of LLMs while managing these constraints effectively. From efficient memory utilization to streamlined parameter fine-tuning methods, we will discuss practical approaches to maximize the efficiency of LLMs without sacrificing performance.

Docker for Machine Learning

Speaker Ryan Daniels, University of Cambridge

Writing research software in Python presents numerous challenges to reproducibility - what version of Python is being used? What about the versions of PyTorch, Scikit Learn or Numpy? Should we use Conda, or venv, or Poetry to manage dependencies and environments? How can we control randomness? Do I have the right version of Cuda Toolkit? In principle, given the same data, and same algorithms and methodology, we should be able to reproduce the results of any given experiment to within an acceptable degree of error. Dealing with the above questions introduces significant problems to reproducing experiments in machine learning. In this talk, I would like to convince you that Docker can help alleviate almost all of these questions. Furthermore, combining Docker, git and GitHub can be a powerful workflow, helping to minimise your tech stack, and declutter your python development experience.

How do you unit test an ML model?

Speaker Wahab Kawafi, University of Bristol

Covering methods such as mock testing, simulation, experiment tracking, and dataset curation. With examples in medicine, chemistry, aerospace engineering, and LLMs.

How to make your machine learning code faster

Speaker Edwin Brown, Research Software Engineering, University of Sheffield

Practical guide to profile machine learning code to find bottlenecks and to remove these bottlenecks.

Nvidia Self-paced training courses and other ML resources

Speaker Denis Battistella, Higher Education and Research, Nvidia

Links to the resources discussed in the presentation:

Acknowledgements

We’d like to say thank you to all the speakers and attendees for making Best Practices in AI Afternoon a great success!

Thank you to Emma and Kate from the Centre for Machine Intelligence (CMI) for all the help with organising the event and the CMI and Nvidia for sponsoring the event.

Photos from the day

Karin Sevegnani presenting Karin Sevegnani presenting “Maximizing Efficiency in Large Language Models: Compute, Memory, and Fine-Tuning”

Ryan Daniels presenting Ryan Daniels presenting “Docker for Machine Learning”

Edwin Brown presenting Edwin Brown presenting “How to make your machine learning code faster”

Bob Turner presenting Bob Turner presenting “From Research Software to Software as a Service”

Denis Battistella presenting Denis Battistella presenting “Nvidia Self-paced Training Courses”

Q & A Panel Q & A Panel with Andy Grant, Edwin Brown, Ryan Daniels, and Christopher Wild

Contact Us

For queries relating to collaborating with the RSE team on projects: rse@sheffield.ac.uk

Information and access to JADE II and Bede.

Join our mailing list so as to be notified when we advertise talks and workshops by subscribing to this Google Group.

Queries regarding free research computing support/guidance should be raised via our Code clinic or directed to the University IT helpdesk.