Canonical
on 4 June 2018
Kubernetes and OpenStack solving AI complexities at scale
Stu Miniman and John Boyer of theCUBE interviewed Stephan Fabel, Director of Ubuntu Product and Development at the OpenStack Summit in Vancouver. Read on for the full interview, and to hear more on Kubernetes, Kubeflow and MicroK8s.
OpenStack has entered a new phase, maturing to a point where those in attendance are serious about open infrastructure, and they are serious about implementing Kubernetes.
Interest in Kubernetes is high, yet, there remains an element of pathfinding that needs to be done on the Kubernetes side, but what needs to made clear is that the technology can work perfectly well with OpenStack.
Stephan Fabel told theCUBE: “What we see a use case for Kubernetes and OpenStack to work perfectly well together, be an extension of each other, possibly also sit next to each other.”
“Having something like Kubernetes container based developer APIs that are providing that orchestration layer are the next thing, and they run just perfectly fine on Canonical OpenStack.”
Kubernetes and containers are also an important method for addressing the problem of scale when it comes to dealing with other industry-wide hot topics – AI and machine learning.
Artificial Intelligence isn’t new. As a subject matter it has been present for decades, but it’s never truly been a realistic possibility until recent technological advancements moved it from being solely a topic of science fiction.
True AI is still in its infancy but some sizeable problems are already being identified. One of them being the issue of scale, and the complexity that will inevitably create.
Smart cities, connected cars, the productization of machine learning applications, and more, spells large scale complexity. Use cases that require vast scale, such as these, creates a requirement for machine learning to be managed – a problem that containers and Kubernetes can solve.
Canonical Founder & CEO Mark Shuttleworth on automating the data centre.
This isn’t just a pipe dream either, Fabel said: “Every single car manufacturer that we’re talking to is interested in solving the operational problem of running machine learning frameworks as a service.” The operational challenges presented by this are difficult to overcome, but customers are seeking solutions from Canonical.
Kubeflow, a machine learning framework that runs on top of Kubernetes, is becoming one of the key enablers of machine learning frameworks as a service – on top of Kubernetes, that can lower the barriers to getting started with this new generation of workloads.
Fabel said: “I think they’re a great example because they can really show how that as a service notion can be implemented on top of a virtualisation platform,” whether that is KVM, bare-metal, or on OpenStack.
MicroK8s
Finding a starting point is often one of the biggest challenges. So lowering the barrier to entry to something like Kubernetes can be vital.
MicroK8s started as a solution to the problem of standing up a Kubernetes cluster on a public cloud when there’s no internet connection, or working on a local laptop. Kubeflow, can also run on MicroK8s, with offerings such as TensorBoard, and the ability to work in a fashion that is compatible to an on-premises and public machine learning framework.
That was just the start and MicroK8s was formed as a solution to offerings in the market that tend to be more cumbersome than developers would like.
Working like a Snap on your machine, you have a Kubernetes API, and in no time at all the user can enable DNS and be up and running.