OpenStack has entered a new phase, maturing to a point where those in attendance are serious about open infrastructure, and they are serious about implementing Kubernetes.
Interest in Kubernetes is high, yet, there remains an element of pathfinding that needs to be done on the Kubernetes side, but what needs to made clear is that the technology can work perfectly well with OpenStack.
Stephan Fabel told theCUBE: “What we see a use case for Kubernetes and OpenStack to work perfectly well together, be an extension of each other, possibly also sit next to each other.”
“Having something like Kubernetes container based developer APIs that are providing that orchestration layer are the next thing, and they run just perfectly fine on Canonical OpenStack.”
Kubernetes and containers are also an important method for addressing the problem of scale when it comes to dealing with other industry-wide hot topics – AI and machine learning.
Artificial Intelligence isn’t new. As a subject matter it has been present for decades, but it’s never truly been a realistic possibility until recent technological advancements moved it from being solely a topic of science fiction.
True AI is still in its infancy but some sizeable problems are already being identified. One of them being the issue of scale, and the complexity that will inevitably create.
Smart cities, connected cars, the productization of machine learning applications, and more, spells large scale complexity. Use cases that require vast scale, such as these, creates a requirement for machine learning to be managed – a problem that containers and Kubernetes can solve.
This isn’t just a pipe dream either, Fabel said: “Every single car manufacturer that we’re talking to is interested in solving the operational problem of running machine learning frameworks as a service.” The operational challenges presented by this are difficult to overcome, but customers are seeking solutions from Canonical.
Kubeflow, a machine learning framework that runs on top of Kubernetes, is becoming one of the key enablers of machine learning frameworks as a service – on top of Kubernetes, that can lower the barriers to getting started with this new generation of workloads.
Fabel said: “I think they’re a great example because they can really show how that as a service notion can be implemented on top of a virtualisation platform,” whether that is KVM, bare-metal, or on OpenStack.
Finding a starting point is often one of the biggest challenges. So lowering the barrier to entry to something like Kubernetes can be vital.
MicroK8s started as a solution to the problem of standing up a Kubernetes cluster on a public cloud when there’s no internet connection, or working on a local laptop. Kubeflow, can also run on MicroK8s, with offerings such as TensorBoard, and the ability to work in a fashion that is compatible to an on-premises and public machine learning framework.
That was just the start and MicroK8s was formed as a solution to offerings in the market that tend to be more cumbersome than developers would like.
Working like a Snap on your machine, you have a Kubernetes API, and in no time at all the user can enable DNS and be up and running.
Ubuntu offers all the training, software infrastructure, tools, services and support you need for your public and private clouds.
For telecommunications companies evaluating ways to transition and modernise their network infrastructure, cloud-native, container-based microservices architectures are a powerful solution in meeting requirements of compute needs through…
The Canonical commissioned study found that 60% of financial services businesses bank on the benefits of multi-cloud to support emerging technologies, but significant skills gap still exists An independent report by 451 Research, ‘Multi-…
Want to enable your developers to deploy highly available web applications, web services or microservices? Do you also need to cater for existing and legacy workloads? Sounds like your requirements are in the sweet spot between what…