OPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS
1
OPERATIONALIZING MACHINE LEARNING USING GPU 1 ACCELERATED, - - PowerPoint PPT Presentation
OPERATIONALIZING MACHINE LEARNING USING GPU 1 ACCELERATED, IN-DATABASE ANALYTICS Why GPUs? Performance Increase A Tale of Numbers Infrastructure Cost Savings 100x 75% Performance Costs 100x gains over traditional 75% reduction in
1
2
A Tale of Numbers
100x 75%
Performance 100x gains over traditional RDBMS / NoSQL / In-Mem Databases Cores Modern GPUs can consist of up to 3000+ cores compared to 32 in a CPU Costs 75% reduction in infrastructure costs, licensing, staff, etc. More with Less Increase performance, throughput, capability while minimizing the costs to support the business
Performance Increase Infrastructure Cost Savings
vs 32
GPUs
3
4
Predict y using function on data x
5
No shortage of techniques and programing languages
6
Python and SQL cover almost all the algorithms in that scary spider and Kinetica supports all Python libraries!
7
process big data: batch and streams
analytics
machine consumers
9
9
SPECIALIZED AI/ DATA SCIENCE TOOLS
SUBSET
DATA SCIENTISTS BUSINESS USERS
EXTRACT EXTRACTING DATA FOR AI IS EXPENSIVE AND SLOW ENTERPRISES STRUGGLE TO MAKE AI MODELS AVAILABLE TO BUSINESS ???
10
Monte Carlo Risk Custom Function 2 Custom Function 3
API EXPOSES CUSTOM FUNCTIONS WHICH CAN BE MADE AVAILABLE TO BUSINESS USERS
BUSINESS USERS DATA SCIENTISTS
UDFs
11
python
12
SQL UDF
python
SQL
Various ETL/ELT
Head Node Worker 1
KINETICA: 10 Node Cluster
Worker 9
Fact and dimensions tables for various Use Cases Billions of rows Massive Stream Ingestion Massive Fast Analytics Apache Tomcat Applications Servers
Full Model Pipeline 1 Various ETL/ELT Full Model Pipeline N Prompts Project
Major U.S Retailer
13
Fast Streaming Projects Fast Analytics Projects
15
A Parametric ModelPython Using TensorFlow Model Training
Table
Model Serving
Model Analytics
UDF train_nd_udf.py
Machine 0 Rank 0 Tom 0
Table mnist_training Shard 0 Table TFModel Shard 0 Table mnist_inference Shard 0 Table mnist_inference_out Shard 0
Tom 1
Table mnist_training Shard 1 Table TFModel Shard 1 Table mnist_inference Shard 1 Table mnist_inference_out Shard 1
Tom 2
Table mnist_training Shard 2 Table TFModel Shard 2 Table mnist_inference Shard 2 Table mnist_inference_out Shard 2
Tom 3
Table mnist_training Shard 3 Table TFModel Shard 3 Table mnist_inference Shard 3 Table mnist_inference_out Shard 3
Machine 0 Rank 0 Tom 4
Table mnist_training Shard 4 Table TFModel Shard 4 Table mnist_inference Shard 4 Table mnist_inference_out Shard 4
Tom 5
Table mnist_training Shard 5 Table TFModel Shard 5 Table mnist_inference Shard 5 Table mnist_inference_out Shard 5
Tom 6
Table mnist_training Shard 6 Table TFModel Shard 6 Table mnist_inference Shard 6 Table mnist_inference_out Shard 6
Tom 7
Table mnist_training Shard 7 Table TFModel Shard 7 Table mnist_inference Shard 7 Table mnist_inference_out Shard 7
UDF UDF UDF UDF UDF UDF UDF UDF
info@kinetica.com
Come get your copy of the O’Reilly Book at Booth G.01!