What we'll do
We’ve all heard that 5G is going to change wireless communication, with incredible bandwidths enabling a new era of connectivity and the realisation of the Internet of Things. (No doubt we’re all looking forward to our grandchildren finally enjoying the internet speeds of the developing world.) With all those devices, there will no doubt be spare computing power for the next generation of citizen science like Folding at Home. Coupled with recent developments in distributed deep learning, perhaps this represents the threshold of a new era of AI. After all, cluster computing and the cloud have been hugely successful in scaling tasks, we can only go upwards. But there’s a problem. Even though 5G promises massive potential increase in transmission speed and volume, communication costs are always the bottleneck in distributed computing.
Enter Rixon Crane, a PhD student from UQ. He sees a world where every mobile phone could be part of a giant supercomputer. Rather than working on making deep learning computationally efficient, he’s building a system that makes deep learning communicationally efficient. By deliberately using expensive local training steps, he’s designed a way to minimise overall worker/server communication rounds, increase the scalability of deep learning training and still ensure efficient convergence. To do this, he uses a technique known as “second order optimisation”, which up until recently had too many limitations to be used in most deep learning optimisation. Plus he came up with a great acronym, which (as we all know) is the hardest part of research and development. Yes, this talk is going to get unavoidably technical, but Rixon has a way with breaking down complex ideas into digestible chunks. So come along, learn about some bleeding edge model training techniques and network with fellow pioneers! This talk promises a vision of the future of distributed computing, and it’s awesome that it’s happening here in Brisbane.