Now Reading
Good News For Data Scientists, Google Launches AI Platform To Democratize ML

Good News For Data Scientists, Google Launches AI Platform To Democratize ML


At Cloud Next’19, an ongoing event at San Francisco, Google launched its very own AI platform. This platform aims at making the life of machine learning developers, data scientists and data engineers easy.

This AI Platform supports Kubeflow, Google’s open-source platform, which lets the users build portable ML pipelines that can be run on-premises or on Google Cloud without significant code changes. And, Kubeflow, AI Hub, and notebooks can be used for no charge.



“So you can start with exploratory data analysis, start to build models using a data scientist, decide you want to use a specific model, and then essentially with one click be able to deploy it in our Google Cloud or in other clouds, or on-premises running on top of Kubernetes,” said Google Cloud chief AI scientist Andrew Moore on the eve of Cloud Next’19 event.

Along with the announcement of its AI platform, Google also made the following announcements:

End-to-end ML Pipeline with AI Platform

Google’s cloud storage or BigQuery to store data while building data labeling service for training. And, then import the labeled data to AutoML to train the model directly.

Once the training is done, users can run their deep learning models via Google Cloud Platform in a serverless environment or using microservices provided by Kubeflow.

AI Platform runs the training in the cloud. Train can be done with a built-in algorithm (beta)on the  dataset without writing a training application. If built-in algorithms do not fit the use case, a training application can be created to run on AI Platform.

The AI Platform training service is designed to have as little impact on the application as possible.

The data in the training job must obey the following rules to run on AI Platform:

  • The data must be in a format that can be read and fed to TensorFlow code.
  • The data must be in a location that code can access.Which means it should be stored with one of the GCP storage or big data services.

The Kubeflow pipelines enables model management and end-to-end work flows control. These pipelines can be leveraged to build reusable end-to-end ML pipelines.

See Also

The AI Platform training service doesn’t provide any special interface for working with GPUs. The user can specify GPU-enabled machines to run the job, and the service allocates it.

With domain expertise from partners like Intel, Cisco and Nvidia, GCP aims to democratise the use of ML on a large scale.

Check more about the AI Platform here


Enjoyed this story? Join our Telegram group. And be part of an engaging community.

Provide your comments below

comments

What's Your Reaction?
Excited
0
Happy
0
In Love
0
Not Sure
0
Silly
0
Scroll To Top