Mobile Data Management Meets Deep Learning Wang-Chien Lee - - PowerPoint PPT Presentation

mobile data management meets deep learning
SMART_READER_LITE
LIVE PREVIEW

Mobile Data Management Meets Deep Learning Wang-Chien Lee - - PowerPoint PPT Presentation

Mobile Data Management Meets Deep Learning Wang-Chien Lee Intelligent Pervasive Data Access ( i PDA) Group Pennsylvania State University wlee@cse.psu.edu 2 MDM June 2019 Vision of Ubiquitous Computing n Ubiquitous computing names the third


slide-1
SLIDE 1

Wang-Chien Lee Intelligent Pervasive Data Access (iPDA) Group Pennsylvania State University

wlee@cse.psu.edu

Mobile Data Management Meets Deep Learning

slide-2
SLIDE 2

June 2019 MDM 2

slide-3
SLIDE 3

Vision of Ubiquitous Computing

n Ubiquitous computing names the third wave in

computing, just now beginning. First were mainframes, each shared by lots of people. Now we are in the personal computing era, person and machine staring uneasily at each other across the

  • desktop. Next comes ubiquitous computing, or the

age of calm technology, when technology recedes into the background of our lives.

  • - by Mark Weiser

n The most profound technologies are those that

  • disappear. They wave themselves into the fabric
  • f everyday life until they are indistinguishable

from it.

3 June 2019 MDM

slide-4
SLIDE 4

4

Party on Friday…

n Update Smart Phone’s calendar

with guests names.

n Make a note to order food from

Dinner-on-Wheels.

n Update shopping list based on the

guests drinking preferences.

n Don’t forget to swipe that last can

  • f beer’s UPC/RFID label.

n The shopping list is always up-to-

date.

June 2019 MDM

slide-5
SLIDE 5

n Approach a local supermarket n AutoPC informs you that you are near a

supermarket

n It informs you the soda and beer are on sale,

and reminds you that your next appointment is in 1 hour.

n There is enough time based on the latest

traffic report.

Party on Friday…

5 June 2019 MDM

slide-6
SLIDE 6

6

Party on Friday…

n TGIF… n Smart Phone reminds you that you need to

  • rder food by noon.

n It downloads the Dinner-on-Wheels menu from

the Web on your PC with the guests’ preferences marked.

n It sends the shopping list to your

CO-OP’s PC.

n Everything will be delivered by the time

you get home in the evening.

June 2019 MDM

slide-7
SLIDE 7

Mobile Data Management

n An important step proceeding the vision of

Ubiquitous computing is mobile computing.

n The system and networking communities have

Mobicom.

n There are needs for a forum to discuss and

address research issues related to data, and

  • ther aspects…

n Prelude: 1998 Workshop on Mobile Data Access

in Singapore.

n Kick Off: 1999 International Conference on

Mobile Data Management in Hong Kong.

June 2019 MDM 7

slide-8
SLIDE 8

MDM Sessions – Early Years

1999 Wireless Networks and Communications Transaction Processing in Mobile Environments Ubiquitous Information Services Mobile Data Replication and Catching Mobility and Location Management 2001 Data Management Architectures Content Delivery Data Broadcasting Caching and Hoarding Coping with Movement Network and System issues 2002 Mobile and Disconnected Operation E-Commerce Data Allocation and Replication Moving Objects Location Management and Awareness

June 2019 MDM 8

slide-9
SLIDE 9

MDM Sessions – In Transition

June 2019 MDM 9

2009 Location Data Management Mobile Peer-to-Peer Networks Embedded Devices and Applications Ad Hoc and Social Networks Sensor and Streaming Data Processing Location Based Services Mobile Data Dissemination and Access Location Privacy and Mining Mobile Peer-to-Peer Networks 2010 Localization and Location-Based Services GIS, Multimedia, and Storage Privacy and Trust Management Query Processing for Location-Based Services Wireless Networks Query Processing in Wireless Sensor Networks Moving Objects 2011 Location-Based Services and Query Optimization Moving Objects and Trajectories Mobility Personalization and Privacy Applications Vehicular and Mobile Networks Wireless Networks Pervasive Computing

slide-10
SLIDE 10

MDM Sessions – Recent Years

June 2019 MDM 10

2016 Information Management on Road Networks Query Processing and Information Search/Retrieval Smart City and Urban Applications Mining and Prediction for Streams and Moving Objects Social Media and Social Networks Ride Sharing, Road Networks and Routes Systems and Platforms Indexing and Querying: Road Networks, Moving Objects, and Trajectories Privacy and Security 2017 Location Services Mobile Data Processing Spatial+X Query Processing Ride Sharing and Recommendations Traffic Data Mining Connected Vehicles Localization and Traffic Analysis Trip Planning Trajectory Mining 2018 Trip Planning Data Mining and Machine Learning on Mobile Data 1 Trajectory Mining Private Query Processing and Ride Sharing Mobile Data Processing Crowd Sourcing and LBSN

slide-11
SLIDE 11

MDM Research Areas

n Essential/Important Issues

l Mobility and Location Management l Application, System and Network Issues l Mobile Data Processing, Query Processing l Privacy and Security

n Disappeared

l Mobile Data Replication, Caching and Hoarding l Content Delivery, Data Broadcasting

n Emerging Topics

l Smart City and Urban Applications, Trip Planning l Mining and Prediction for Streams and Moving Objects l Trajectory Mining, Traffic Data Mining, Ride Sharing

and Recommendations June 2019 MDM 11

slide-12
SLIDE 12

Ubiquitous Comp – Step Forward

n We are moving further towards the vision of

Ubiquitous Computing

l Abundant communication bandwidth l Abundant computing power

n Computing is becoming Invisible

l Smart city, Smart building, Smart Vehicles l Smart watch, Smart Speakers, Smart applications

n We are in a process of smartening all the

encounters in our daily life

l Enabled by abundant data and machine learning,

especially with the timely breakthrough of deep learning technology June 2019 MDM 12

slide-13
SLIDE 13

Breakthroughs of Deep Learning

n In 2012, AlexNet achieved 16% error rate in image

classification on ImageNet. Then, VGG, GoogleNet, ResNet further improves to 7.3%, 6.7%, 3.5% compared with human average error 5%.

n In 2014, DeepFace identifies faces with 97.35%

accuracy, competitive with human performance.

n In 2016, AlphaGo defeats a World Champ Lee

Sedol (4:1) and is awarded an honorary 9-dan title.

n Models are proposed to various NLP apps, e.g.,

Word2Vec, Seq2Seq, Transformer. In 2018, BERT

  • btains state-of-the-art results on 11 NLP tasks,

described as the “Imagenet moment for NLP”.

June 2019 MDM 13

slide-14
SLIDE 14

June 2019 MDM 14

slide-15
SLIDE 15

Potential Research

n Location Based Social Networks

l Network representation learning

n Trajectory Mining

l Trajectory representation learning l Travel time estimation

n Intelligent Transportation Systems

l Traffic Incident Inference l Traffic forecast l Traffic Sign Recognition

June 2019 MDM 15

slide-16
SLIDE 16

Location-Based Social Networks

June 2019 MDM 16

slide-17
SLIDE 17

Yelp Dataset

1M users

  • user_id, name, review_count,

yelping_since, friends, useful, funny, cool, fans, elite, average_stars, compliment_hot, compliment_more, compliment_profile, compliment_cute, compliment_list, compliment_note, compliment_plain, compliment_cool,compliment_funny, compliment_writer, compliment_photos

946K tips

  • user_id, business_id, text, likes

144K restaurants

  • business_id, name, neighborhood,

address, city, state, postal_code, lng, lat, stars, review_count, is_open, attributes: [parking, payments, ...], categories: [tags], hours

125K check-ins

  • business_id, time: [(time, count)]

4.1M reviews

  • review_id, user_id, business_id, star,

date, text, useful, funny, cool

17

slide-18
SLIDE 18

Functionality

n Restaurant search:

l Given a restaurant, recommend similar restaurants l Formulate as k-nearest neighbor (KNN) search problem

n Personalized restaurant recommendation:

l Given a user, recommend restaurants of her interests l Formulate as a link prediction problem

n Restaurant categorization:

l Given a restaurant, classify it into categories l Formulate as a classification problem

n Friendship recommendation:

l Given a user, recommend new friends to her. l Formulated as a similarity search problem

June 2019 MDM 18

?

?

slide-19
SLIDE 19

Data Mining on Network Data

Many applications of location based social network data and service functionality are formulated as classical data mining tasks:

n Node classification

l Predict the type of a given node

n Link prediction

l Predict whether two nodes are linked

n Clustering/Community detection

l Identify densely linked clusters of nodes

n Similarity search

l How similar/relevant are two nodes? l How similar are two (sub)networks

MDM 19 June 2019

slide-20
SLIDE 20

n Network data analytics often involve prediction tasks

  • ver nodes/edges. To achieve good performance,

feature engineering is essential but labor-intensive.

n Open problem: Efficient and automatic feature

learning

l Ideally, the learned features are task-independent!

Automatic Feature Engineering

20 MDM June 2019

Feature Engineering

slide-21
SLIDE 21

HIN2Vec (Fu et al, CIKM’17)

n To support a variety of LBSN applications, HIN2Vec

automatically generates latent embeddings with inherent properties to serve as input features.

n HIN2Vec considers heterogeneous data n HIN2Vec distinguishes the different relationships

between nodes, and thus preserves more precise information

n HIN2Vec learns meaningful representations by

encoding the rich information embedded in meta- paths and network structure.

l Nodes with strong relationships are close to each other. l Relationship vectors provide analytical insights

June 2019 MDM 21

slide-22
SLIDE 22

HIN2Vec Framework

June 2019 MDM 22

HIN2Vec Phase I Training data preparation

random walk, negative sampling

Phase II Representation learning node vectors Wx training set targeted meta- paths meta-path vectors WR r x y

WX WY f01(WR )

Restaurant search (K Nearest Neighbors) Personalized restaurant Recommendation (Link Prediction) ? Restaurant categorization (Node Classification) ? Friendship recommendation (Similarity Search)

?

Canyon pizza College pizza Five guys piz za fri e s

slide-23
SLIDE 23

Trajectory Mining

n Many trajectory

datasets made available publicly.

n Applications

l Search for similar

trajectories

l Trajectory clustering l Travel time estimation

n Learned trajectory

representations may be used for some applications.

June 2019 MDM 23

Porto taxi data, Taxi Service Trajectory Prediction Challenge@ ECML/PKDD 2015, contains 1.7 million taxi trajectories of 442 taxis in Porto, Portugal over 19 months.

slide-24
SLIDE 24

Trajectory Representation Learning

n Trajectory Clustering

l To learn trajectory embeddings by capturing mobile

users’ moving behaviors for trajectory clustering applications.

l Yao, Di, et al. Trajectory clustering via deep

representation learning, 2017 international joint conference on neural networks (IJCNN), 2017

n Trajectory Similarity Computation

l To learn trajectory embeddings by capturing mobile

users’ moving behaviors for trajectory similarity computation.

l X. Li, et al., Deep Representation Learning for

Trajectory Similarity Computation, International Conference on Data Engineering (ICDE). 2018. June 2019 MDM 24

slide-25
SLIDE 25

Trajectory2Vec

n Trajectory Preprocessing Layer

l It applies existing techniques for data cleaning by

filtering low-quality sample points

n Moving Behavior Feature Extraction Layer

l It applies a sliding window to transform a raw trajectory

as a sequence of windows containing sample points.

l Generate a number of features (e.g., time interval,

moving distance, change of speed, etc) for each window. June 2019 MDM 25

x1 x2 x3 x4 x5 x6 x7 trajectory windows b1 features w1 w2 w3 w4 b3 b2 b4

slide-26
SLIDE 26

Seq2Seq Auto-encoder

n It applies Seq2Seq model to encode a trajectory

(transformed as B={b1, b1, ...}) into a low- dimensional vectors which in turn is decoded back to the original B.

June 2019 MDM 26

Decoder Encoder b1 b2 bn ... b1 b1 b2 ... bn-1

bn

h0 Trajectory embeddings Learned by minimizing the re-construction error

slide-27
SLIDE 27

T2Vec - Data preprocessing

n For low sampling rate

l For a trajectory T, t2vec splits it interleavingly to Ta

and Tb (like downsampling)

l Then, the proposed RNN-based encoder-decoder

aims to encode Ta into a low-dimensional vector which is used to decode Tb

n For noisy data

l It randomly adds more noises to sample data

June 2019 MDM 27

x1 x2 x3 x4 x5 x6 x7

Ta Tb

slide-28
SLIDE 28

T2Vwec - Seq2Seq Auto-encoder

n Apply Seq2Seq model to encode Ta into a low-

dimensional vector and then decode in turn to Tb

June 2019 MDM 28

Decoder Encoder x1 x3 xn ... x2 x2 x4 ... xn-3

xn+1

h0 Trajectory embeddings Learned by minimizing the re-construction error

slide-29
SLIDE 29

Travel Time Estimation

n Applications: Route planning,

Navigation, Ridesharing and Traffic dispatching, etc.

l H. Zhang, et al., Deeptravel: a

neural network based travel time estimation model with auxiliary supervision, International Joint Conference on Artificial Intelligence (IJCAI-18).

l D. Wang, et al., When Will You

Arrive? Estimating Travel Time Based on Deep Neural Networks, AAAI Conference on Artificial Intelligence (AAAI-18). June 2019 MDM 29

slide-30
SLIDE 30

DeepTravel – Feature Extraction

n Partition a trajectory into a grid and map each

GPS sample point into a grid cell.

n Extract features for each cell, including spatial

and temporal embeddings, driving state features, short-term and long-term traffic features.

June 2019 MDM 30

… … … … … … …

Spatial Embedding Temporal Embedding Driving State Features Short-term Traffic Features Long-term Traffic Features

Feather Representation Lays

slide-31
SLIDE 31

DeepTravel – Prediction

n The prediction layer consists of two parts.

l BiLSTM: uses the extracted features to infer travel time l Dual loss: forces the model to learn by simultaneously

predicting forward interval from the start point and backward interval from the destination to each intermediate GPS sample point. June 2019 MDM 31

… …

(ℎ1 + ⋯ + ℎ& − 2 + ℎ& − 1 + ℎ&) (ℎ& + 1 + ℎ& + 2 + ⋯ + ℎ*)

Forward Interval Loss Predict forward Interval

FC layer

Backward Interval Loss Predict backward Interval

FC layer

slide-32
SLIDE 32

DeepTTE – Model Architecture

June 2019 MDM 32

Figure from DeepTTE paper

slide-33
SLIDE 33

DeepTTE – Geo-Convolution

June 2019 MDM 33

… .. . .. . …

GPS trajectory Geo-Conv with multiple kernels 16 channel features K=3 Filter B K=3 Filter B Filter B K=3 K=3 Filter B K=3 Filter B K=3 Filter B K=3 Filter B K=3 Filter B K=3 Filter B K=3 Filter B Filter A K=3 ... … i-th local path

… … … …

Distance Concatenate features and distance

slide-34
SLIDE 34

Intelligent Transportation Systems

n Traffic Incident Inference n Traffic Forecast n Traffic Sign Recognition

June 2019 MDM 34

slide-35
SLIDE 35

Traffic Incident Inference

n Estimate traffic accident risk by mining big and

heterogeneous data.

n Q. Chen, et al., Learning deep representation

from big and heterogeneous data for traffic accident inference’. AAAI, Toronto, Canada, 2016, pp. 338–344.

June 2019 MDM 35 Of all the systems with which people have to deal every day, road traffic systems are the most complex and dangerous.

World report on road traffic injury prevention, published by World Health Organization 2004.

slide-36
SLIDE 36

Accident Prediction Framework

June 2019 MDM 36

The output of the SdAE is the latent representation of human mobility in each grid cell, which serve as input features to the classifier to predict accident risk

slide-37
SLIDE 37

Traffic Forecast

n Predicting the most likely traffic measurements, e.g.,

speed, traffic flow, in the next H time steps, given previous M traffic observations.

n B. Yu, et al., Spatio-temporal graph convolutional

neural network: a deep learning framework for traffic forecasting, International Joint Conference on Artificial Intelligence (IJCAI-18)

June 2019 MDM 37

slide-38
SLIDE 38

Graph-Structured Traffic Data

n To fully utilize spatial information, the traffic network

is modeled by a general graph

n Temporal patterns of traffic flows are also important

June 2019 MDM 38 Time

!" !"#$ !"%$

… … !" = ((

", *, +)

(

": monitor stations

*: connectedness +: edge weights

slide-39
SLIDE 39

STGCN Framework

n ST-Conv Block consists of:

l Graph Convolutional Network

(GCN) extracts meaningful spatial patterns and features

l Gated CNNs captures temporal

dynamic behaviors of traffic flow June 2019 MDM 39

Temporal Gated-Conv Spatial Graph-Conv Temporal Gated-Conv

!" !"#$ !"%$ !′" !′"#$ !′"%$

ST-Conv Block

'

ST-Conv Block ST-Conv Block Output Layer

Gt-M+1 Gt

v* Spatio-Temporal Graph Convolutional Networks

slide-40
SLIDE 40

Traffic Sign Recognition

n Unmanned vehicle has

attracted significant attention.

n Traffic sign recognition is an

essential functionality for the upcoming unmanned vehicles.

n P. Sermanet and Y. LeCun, Traffic sign recognition

with multi-scale convolutional networks, International Joint Conference on Neural Networks (IJCNN), 2011, pp. 2809–2813.

n J. Zhang, et al., A shallow network with combined

pooling for fast traffic sign recognition. Information 8(2), 2017.

40 June 2019 MDM

slide-41
SLIDE 41

GTSRB Competition

n The German Traffic Sign Recognition Benchmark n The German Traffic Sign Detection Benchmark n Challenging Examples

June 2019 MDM 41

slide-42
SLIDE 42

Two-stage ConvNet

42

Convolutions Subsampling Output Full Connection Subsampling Convolutions Convolutions 1st Stage 2nd Stage Classifier Input

The outputs of all the stages are fed to the classifier. This allows the classifier to use not just high-level features but also pooled low-level features, which tend to be more local, less invariant, and more accurately encode local motifs. The 2nd stage extracts “global” and invariant shapes and structures. The first stage extracts “local” motifs with more precise details.

June 2019 MDM

slide-43
SLIDE 43

Three-Stage Shallow CNNs

43

n Each stage includes a convolutional layer and a

subsampling layer

Convolutions Pooling/ Subsampling Output Pooling/ Subsampling Softmax_loss Convolutions 1st Stage 2nd Stage Input 3rd Stage Full Connect Mean removal, whiten, convolutions Pooling Subsampling The second full-connected layer is similar to a single-hidden layer feedforward neural network (SLFN), and the output size is equal to the class number. The first fully connected layer is identical to a convolutional layer, aimed at reducing the dimensionality and preparing for the classification.

June 2019 MDM

slide-44
SLIDE 44

Conclusion

n We are moving a step further towards the vision

  • f Ubiquitous Computing.

n Research in MDM has expanded from accessing

mobile data, managing mobile data, to now smartening mobile applications.

n Recent breakthrough in deep learning technology

brings opportunities and promises to many MDM research areas.

n Look forward to seeing blossom of

research in the coming decade, when we celebrate the 30th Anniversary

  • f MDM.

June 2019 MDM 44