SlideShare a Scribd company logo
Learning Deep
Broadband Network@HOME
Hongjoo LEE
Who am I?
● Machine Learning Engineer
○ Fraud Detection System
○ Software Defect Prediction
● Software Engineer
○ Email Services (40+ mil. users)
○ High traffic server (IPC, network, concurrent programming)
● MPhil, HKUST
○ Major : Software Engineering based on ML tech
○ Research interests : ML, NLP, IR
Outline
Data Collection Time series Analysis Forecast Modeling Anomaly Detection
Naive approach
Logging SpeedTest
Data preparation
Handling time series
Seasonal Trend Decomposition Rolling Forecast Basic approaches
Stationarity
Autoregression, Moving Average
Autocorrelation
ARIMA Multivariate Gaussian
LSTM
Home Network
Home Network
Home Network
Anomaly Detection (Naive approach in 2015)
Problem definition
● Detect abnormal states of Home Network
● Anomaly detection for time series
○ Finding outlier data points relative to some usual signal
Types of anomalies in time series
● Additive outliers
Types of anomalies in time series
● Temporal changes
Types of anomalies in time series
● Level shift
Outline
Data Collection Time series Analysis Forecast Modeling Anomaly Detection
Naive approach
Logging SpeedTest
Data preparation
Handling time series
Seasonal Trend Decomposition Rolling Forecast Basic approaches
Stationarity
Autoregression, Moving Average
Autocorrelation
ARIMA Multivariate Gaussian
LSTM
Logging Data
● Speedtest-cli
● Every 5 minutes for 3 Month. ⇒ 20k observations.
$ speedtest-cli --simple
Ping: 35.811 ms
Download: 68.08 Mbit/s
Upload: 19.43 Mbit/s
$ crontab -l
*/5 * * * * echo ‘>>> ‘$(date) >> $LOGFILE; speedtest-cli --simple >> $LOGFILE
2>&1
Logging Data
● Log output
$ more $LOGFILE
>>> Thu Apr 13 10:35:01 KST 2017
Ping: 42.978 ms
Download: 47.61 Mbit/s
Upload: 18.97 Mbit/s
>>> Thu Apr 13 10:40:01 KST 2017
Ping: 103.57 ms
Download: 33.11 Mbit/s
Upload: 18.95 Mbit/s
>>> Thu Apr 13 10:45:01 KST 2017
Ping: 47.668 ms
Download: 54.14 Mbit/s
Upload: 4.01 Mbit/s
Data preparation
● Parse data
class SpeedTest(object):
def __init__(self, string):
self.__string = string
self.__pos = 0
self.datetime = None# for DatetimeIndex
self.ping = None # ping test in ms
self.download = None# down speed in Mbit/sec
self.upload = None # up speed in Mbit/sec
def __iter__(self):
return self
def next(self):
…
Data preparation
● Build panda DataFrame
speedtests = [st for st in SpeedTests(logstring)]
dt_index = pd.date_range(
speedtests[0].datetime.replace(second=0, microsecond=0),
periods=len(speedtests), freq='5min')
df = pd.DataFrame(index=dt_index,
data=([st.ping, st.download, st.upload] for st in speedtests),
columns=['ping','down','up'])
Data preparation
● Plot raw data
Data preparation
● Structural breaks
○ Accidental missings for a long period
Data preparation
● Handling missing data
○ Only a few occasional cases
Handling time series
● By DatetimeIndex
○ df[‘2017-04’:’2017-06’]
○ df[‘2017-04’:]
○ df[‘2017-04-01 00:00:00’:]
○ df[df.index.weekday_name == ‘Monday’]
○ df[df.index.minute == 0]
● By TimeGrouper
○ df.groupby(pd.TimeGrouper(‘D’))
○ df.groupby(pd.TimeGrouper(‘M’))
Patterns in time series
● Is there a pattern in 24 hours?
Patterns in time series
● Is there a daily pattern?
Components of Time series data
● Trend :The increasing or decreasing direction in the series.
● Seasonality : The repeating in a period in the series.
● Noise : The random variation in the series.
Components of Time series data
● A time series is a combination of these components.
○ yt
= Tt
+ St
+ Nt
(additive model)
○ yt
= Tt
× St
× Nt
(multiplicative model)
Seasonal Trend Decomposition
from statsmodels.tsa.seasonal import seasonal_decompose
decomposition = seasonal_decompose(week_dn_ts)
plt.plot(week_dn_ts) # Original
plt.plot(decomposition.seasonal)
plt.plot(decomposition.trend)
Rolling Forecast
A B C
Rolling Forecast
from statsmodels.tsa.arima_model import ARIMA
forecasts = list()
history = [x for x in train_X]
for t in range(len(test_X)): # for each new observation
model = ARIMA(history, order=order) # update the model
y_hat = model.fit().forecast() # forecast one step ahead
forecasts.append(y_hat) # store predictions
history.append(test_X[t]) # keep history updated
Residuals ~ N( , 2
)
residuals = [test[t] - forecasts[t] for t in range(len(test_X))]
residuals = pd.DataFrame(residuals)
residuals.plot(kind=’kde’)
Anomaly Detection (Basic approach)
● IQR (Inter Quartile Range)
● 2-5 Standard Deviation
● MAD (Median Absolute Deviation)
Anomaly Detection (Naive approach)
● Inter Quartile Range
Anomaly Detection (Naive approach)
● Inter Quartile Range
○ NumPy
○ Pandas
q1, q3 = np.percentile(col, [25, 75])
iqr = q3 - q1
np.where((col < q1 - 1.5*iqr) | (col > q3 + 1.5*iqr))
q1 = df[‘col’].quantile(.25)
q3 = df[‘col’].quantile(.75)
iqr = q3 - q1
df.loc[~df[‘col’].between(q1-1.5*iqr, q3+1.5*iqr),’col’]
Anomaly Detection (Naive approach)
● 2-5 Standard Deviation
Anomaly Detection (Naive approach)
● 2-5 Standard Deviation
○ NumPy
○ Pandas
std = pd[‘col’].std()
med = pd[‘col’].median()
df.loc[~df[‘col’].between(med - 3*std, med + 3*std), 0]
std = np.std(col)
med = np.median(col)
np.where((col < med - 3*std) | (col < med + 3*std))
Anomaly Detection (Naive approach)
● MAD (Median Absolute Deviation)
○ MAD = median(|Xi
- median(X)|)
○ “Detecting outliers: Do not use standard deviation around the mean, use absolute deviation
around the median” - Christopher Leys (2013)
Outline
Data Collection Time series Analysis Forecast Modeling Anomaly Detection
Naive approach
Logging SpeedTest
Data preparation
Handling time series
Seasonal Trend Decomposition Rolling Forecast Basic approaches
Stationarity
Autoregression, Moving Average
Autocorrelation
ARIMA Multivariate Gaussian
LSTM
Stationary Series Criterion
● The mean, variance and covariance of the series are time invariant.
stationary non-stationary
Stationary Series Criterion
● The mean, variance and covariance of the series are time invariant.
stationary non-stationary
Stationary Series Criterion
● The mean, variance and covariance of the series are time invariant.
stationary non-stationary
Test Stationarity
Differencing
● A non-stationary series can be made stationary after differencing.
● Instead of modelling the level, we model the change
● Instead of forecasting the level, we forecast the change
● I(d) = yt
- yt-d
● AR + I + MA
Autoregression (AR)
● Autoregression means developing a linear model that uses observations at
previous time steps to predict observations at future time step.
● Because the regression model uses data from the same input variable at
previous time steps, it is referred to as an autoregression
Moving Average (MA)
● MA models look similar to the AR component, but it's dealing with different
values.
● The model account for the possibility of a relationship between a variable
and the residuals from previous periods.
ARIMA(p, d, q)
● Autoregressive Integrated Moving Average
○ AR : A model that uses dependent relationship between an observation and some number of
lagged observations.
○ I : The use of differencing of raw observations in order to make the time series stationary.
○ MA : A model that uses the dependency between an observation and a residual error from a
MA model.
● parameters of ARIMA model
○ p : The number of lag observations included in the model
○ d : the degree of differencing, the number of times that raw observations are differenced
○ q : The size of moving average window.
Identification of ARIMA
● Autocorrelation function(ACF) : measured by a simple correlation between
current observation Yt
and the observation p lags from the current one Yt-p
.
● Partial Autocorrelation Function (PACF) : measured by the degree of
association between Yt
and Yt-p
when the effects at other intermediate time
lags between Yt
and Yt-p
are removed.
● Inference from ACF and PACF : theoretical ACFs and PACFs are available for
various values of the lags of AR and MA components. Therefore, plotting
ACFs and PACFs versus lags and comparing leads to the selection of the
appropriate parameter p and q for ARIMA model
Identification of ARIMA (easy case)
● General characteristics of theoretical ACFs and PACFs
● Reference :
○ http://people.duke.edu/~rnau/411arim3.htm
○ Prof. Robert Nau
model ACF PACF
AR(p) Tail off; Spikes decay towards zero Spikes cutoff to zero after lag p
MA(q) Spikes cutoff to zero after lag q Tails off; Spikes decay towards zero
ARMA(p,q) Tails off; Spikes decay towards zero Tails off; Spikes decay towards zero
Identification of ARIMA (easy case)
Identification of ARIMA (complicated)
Anomaly Detection (Parameter Estimation)
xdown
xup
xdown
xup
Anomaly Detection (Multivariate Gaussian Distribution)
Anomaly Detection (Multivariate Gaussian)
import numpy as np
from scipy.stats import multivariate_normal
def estimate_gaussian(dataset):
mu = np.mean(dataset, axis=0)
sigma = np.cov(dataset.T)
return mu, sigma
def multivariate_gaussian(dataset, mu, sigma):
p = multivariate_normal(mean=mu, cov=sigma)
return p.pdf(dataset)
mu, sigma = estimate_gaussian(train_X)
p = multivariate_gaussian(train_X, mu, sigma)
anomalies = np.where(p < ep) # ep : threshold
Anomaly Detection (Multivariate Gaussian)
import numpy as np
from scipy.stats import multivariate_normal
def estimate_gaussian(dataset):
mu = np.mean(dataset, axis=0)
sigma = np.cov(dataset.T)
return mu, sigma
def multivariate_gaussian(dataset, mu, sigma):
p = multivariate_normal(mean=mu, cov=sigma)
return p.pdf(dataset)
mu, sigma = estimate_gaussian(train_X)
p = multivariate_gaussian(train_X, mu, sigma)
anomalies = np.where(p < ep) # ep : threshold
Anomaly Detection (Multivariate Gaussian)
import numpy as np
from scipy.stats import multivariate_normal
def estimate_gaussian(dataset):
mu = np.mean(dataset, axis=0)
sigma = np.cov(dataset.T)
return mu, sigma
def multivariate_gaussian(dataset, mu, sigma):
p = multivariate_normal(mean=mu, cov=sigma)
return p.pdf(dataset)
mu, sigma = estimate_gaussian(train_X)
p = multivariate_gaussian(train_X, mu, sigma)
anomalies = np.where(p < ep) # ep : threshold
Outline
Data Collection Time series Analysis Forecast Modeling Anomaly Detection
Naive approach
Logging SpeedTest
Data preparation
Handling time series
Seasonal Trend Decomposition Rolling Forecast Basic approaches
Stationarity
Autoregression, Moving Average
Autocorrelation
ARIMA Multivariate Gaussian
LSTM
Long Short-Term Memory
h0
h1
h2
ht-2
ht-1
c0
c1
c2
ct-2
ct-1
x0
x1
x2
xt-2
xt-1
xt
LSTM layer
Long Short-Term Memory
x0
dn
x0
up
x0
pg
xt
dn
x1
up
x2
pg
xt-1
up
xt-1
dn
xt-1
pg
h0
h1
h2
ht-2
ht-1
c0
c1
c2
ct-2
ct-1
x0
x1
x2
xt-2
xt-1
xt
LSTM layer
Long Short-Term Memory
from keras.models import Sequential
from keras.layers import Dense
from keras.layers import LSTM
from sklearn.metrics import mean_squared_error
model = Sequential()
model.add(LSTM(num_neurons, stateful=True, return_sequences=True,
batch_input_shape=(batch_size, timesteps, input_dimension))
model.add(LSTM(num_neurons, stateful=True,
batch_input_shape=(batch_size, timesteps, input_dimension))
model.add(Dense(1))
model.compile(loss='mean_squared_error', optimizer='adam')
for i in range(num_epoch):
model.fit(train_X, y, epochs=1, batch_size=batch_size, shuffle=False)
model.reset_states()
Long Short-Term Memory
● Will allow to model sophisticated and seasonal dependencies in time series
● Very helpful with multiple time series
● On going research, requires a lot of work to build model for time series
Summary
● Be prepared before calling engineers for service failures
● Pythonista has all the powerful tools
○ pandas is great for handling time series
○ statsmodels for analyzing and modeling time series
○ sklearn is such a multi-tool in data science
○ keras is good to start deep learning
● Pythonista needs to understand a few concepts before using the tools
○ Stationarity in time series
○ Autoregressive and Moving Average
○ Means of forecasting, anomaly detection
● Deep Learning for forecasting time series
○ still on-going research
● Do try this at home
Contacts
lee.hongjoo@yandex.com
linkedin.com/in/hongjoo-lee

More Related Content

What's hot (20)

PDF
Mining Frequent Closed Graphs on Evolving Data Streams
Albert Bifet
 
PDF
VAE-type Deep Generative Models
Kenta Oono
 
PDF
Internet of Things Data Science
Albert Bifet
 
PDF
Gradient Estimation Using Stochastic Computation Graphs
Yoonho Lee
 
PDF
Deep learning for molecules, introduction to chainer chemistry
Kenta Oono
 
ZIP
Analysis
Sri Prasanna
 
PDF
Mining Adaptively Frequent Closed Unlabeled Rooted Trees in Data Streams
Albert Bifet
 
PDF
Recursive algorithms
subhashchandra197
 
PDF
Large scale logistic regression and linear support vector machines using spark
Mila, Université de Montréal
 
PDF
02 analysis
Rick Boardman
 
PDF
Scalable Link Discovery for Modern Data-Driven Applications
Holistic Benchmarking of Big Linked Data
 
PPTX
STRIP: stream learning of influence probabilities.
Albert Bifet
 
PPTX
Scaling out logistic regression with Spark
Barak Gitsis
 
PDF
Multinomial Logistic Regression with Apache Spark
DB Tsai
 
PDF
ICML2013読み会 Large-Scale Learning with Less RAM via Randomization
Hidekazu Oiwa
 
PPTX
Aaex4 group2(中英夾雜)
Shiang-Yun Yang
 
PDF
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Universitat Politècnica de Catalunya
 
PPTX
Exploring Optimization in Vowpal Wabbit
Shiladitya Sen
 
PDF
2014-06-20 Multinomial Logistic Regression with Apache Spark
DB Tsai
 
PDF
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
Universitat Politècnica de Catalunya
 
Mining Frequent Closed Graphs on Evolving Data Streams
Albert Bifet
 
VAE-type Deep Generative Models
Kenta Oono
 
Internet of Things Data Science
Albert Bifet
 
Gradient Estimation Using Stochastic Computation Graphs
Yoonho Lee
 
Deep learning for molecules, introduction to chainer chemistry
Kenta Oono
 
Analysis
Sri Prasanna
 
Mining Adaptively Frequent Closed Unlabeled Rooted Trees in Data Streams
Albert Bifet
 
Recursive algorithms
subhashchandra197
 
Large scale logistic regression and linear support vector machines using spark
Mila, Université de Montréal
 
02 analysis
Rick Boardman
 
Scalable Link Discovery for Modern Data-Driven Applications
Holistic Benchmarking of Big Linked Data
 
STRIP: stream learning of influence probabilities.
Albert Bifet
 
Scaling out logistic regression with Spark
Barak Gitsis
 
Multinomial Logistic Regression with Apache Spark
DB Tsai
 
ICML2013読み会 Large-Scale Learning with Less RAM via Randomization
Hidekazu Oiwa
 
Aaex4 group2(中英夾雜)
Shiang-Yun Yang
 
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Universitat Politècnica de Catalunya
 
Exploring Optimization in Vowpal Wabbit
Shiladitya Sen
 
2014-06-20 Multinomial Logistic Regression with Apache Spark
DB Tsai
 
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
Universitat Politècnica de Catalunya
 

Viewers also liked (20)

PDF
Mock it right! A beginner’s guide to world of tests and mocks, Maciej Polańczyk
Pôle Systematic Paris-Region
 
PDF
How to apply deep learning to 3 d objects
Ogushi Masaya
 
PPTX
Type Annotations in Python: Whats, Whys and Wows!
Andreas Dewes
 
PPTX
EuroPython 2017 - How to make money with your Python open-source project
Max Tepkeev
 
PDF
OpenAPI development with Python
Takuro Wada
 
PDF
Analytics Academy 2017 Presentation Slides
HarvardComms
 
PPSX
Fixing The Sales Forecast
SmartFunnel
 
PPTX
Systematically Improving Sales Forecast Accuracy
Inflexion-Point Strategy Partners
 
PDF
IDATE DigiWorld - FTTH global perspective 241017 - Roland Montagne
IDATE DigiWorld
 
PPT
The Future of Social Networks on the Internet: The Need for Semantics
John Breslin
 
PDF
Sales forecast
Casey Robertson
 
PPTX
DigiWorld Future Paris-Bernard Ourghanlian-CTO & CS0- Microsoft
IDATE DigiWorld
 
PDF
Predictive Analytics in Telecommunication
Rising Media Ltd.
 
PDF
Improving Forecast Accuracy
Onur Sezgin
 
PDF
IDATE DigiWorld - FTTH global perspective 241017 Gigabit VF - Roland Montagne
IDATE DigiWorld
 
PDF
A Test of B2B Sales Forecasting Methods
SalesClic
 
PDF
The State of Broadband: Broadband catalyzing sustainable development. Septemb...
Andrés Rodríguez Seijo
 
PPT
Broadband 101 Feasibility Studies
Ann Treacy
 
PDF
IDATE DigiWorld -FTTx markets public - Roland MONTAGNE
IDATE DigiWorld
 
PDF
Understanding RF Fundamentals and the Radio Design of Wireless Networks
Cisco Mobility
 
Mock it right! A beginner’s guide to world of tests and mocks, Maciej Polańczyk
Pôle Systematic Paris-Region
 
How to apply deep learning to 3 d objects
Ogushi Masaya
 
Type Annotations in Python: Whats, Whys and Wows!
Andreas Dewes
 
EuroPython 2017 - How to make money with your Python open-source project
Max Tepkeev
 
OpenAPI development with Python
Takuro Wada
 
Analytics Academy 2017 Presentation Slides
HarvardComms
 
Fixing The Sales Forecast
SmartFunnel
 
Systematically Improving Sales Forecast Accuracy
Inflexion-Point Strategy Partners
 
IDATE DigiWorld - FTTH global perspective 241017 - Roland Montagne
IDATE DigiWorld
 
The Future of Social Networks on the Internet: The Need for Semantics
John Breslin
 
Sales forecast
Casey Robertson
 
DigiWorld Future Paris-Bernard Ourghanlian-CTO & CS0- Microsoft
IDATE DigiWorld
 
Predictive Analytics in Telecommunication
Rising Media Ltd.
 
Improving Forecast Accuracy
Onur Sezgin
 
IDATE DigiWorld - FTTH global perspective 241017 Gigabit VF - Roland Montagne
IDATE DigiWorld
 
A Test of B2B Sales Forecasting Methods
SalesClic
 
The State of Broadband: Broadband catalyzing sustainable development. Septemb...
Andrés Rodríguez Seijo
 
Broadband 101 Feasibility Studies
Ann Treacy
 
IDATE DigiWorld -FTTx markets public - Roland MONTAGNE
IDATE DigiWorld
 
Understanding RF Fundamentals and the Radio Design of Wireless Networks
Cisco Mobility
 
Ad

Similar to EuroPython 2017 - PyData - Deep Learning your Broadband Network @ HOME (20)

PDF
timeseries cheat sheet with example code for R
derekjohnson549253
 
PDF
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Cynthia Freeman
 
PDF
Forecasting time series powerful and simple
Ivo Andreev
 
PPTX
Presentation On Time Series Analysis in Mechine Learning
mahfuzur32785
 
PPTX
Time Series Anomaly Detection with .net and Azure
Marco Parenzan
 
PDF
MFx_Module_3_Properties_of_Time_Series.pdf
tilfani
 
PPTX
Seasonal Decomposition of Time Series Data
Programming Homework Help
 
PDF
Outlier analysis for Temporal Datasets
QuantUniversity
 
PDF
time_series.pdf
PranaySawant20
 
PPTX
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Simplilearn
 
PPTX
Project time series ppt
amar patil
 
PPTX
Time series Modelling Basics
Ashutosh Kumar
 
PPTX
ARIMA model predicts futture values based on past values
aravindhanb13
 
PPTX
Simple math for anomaly detection toufic boubez - metafor software - monito...
tboubez
 
PPTX
Forecasting_CO2_Emissions.pptx
MOINDALVS
 
PPTX
ARIMA.pptx
Ramakrishna Reddy Bijjam
 
PDF
a brief introduction to Arima
Yue Xiangnan
 
PPTX
Time series analysis 101
Balvinder Hira
 
PPTX
Long Memory presentation to SURF
Richard Hunt
 
PDF
Different Models Used In Time Series - InsideAIML
VijaySharma802
 
timeseries cheat sheet with example code for R
derekjohnson549253
 
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Cynthia Freeman
 
Forecasting time series powerful and simple
Ivo Andreev
 
Presentation On Time Series Analysis in Mechine Learning
mahfuzur32785
 
Time Series Anomaly Detection with .net and Azure
Marco Parenzan
 
MFx_Module_3_Properties_of_Time_Series.pdf
tilfani
 
Seasonal Decomposition of Time Series Data
Programming Homework Help
 
Outlier analysis for Temporal Datasets
QuantUniversity
 
time_series.pdf
PranaySawant20
 
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Simplilearn
 
Project time series ppt
amar patil
 
Time series Modelling Basics
Ashutosh Kumar
 
ARIMA model predicts futture values based on past values
aravindhanb13
 
Simple math for anomaly detection toufic boubez - metafor software - monito...
tboubez
 
Forecasting_CO2_Emissions.pptx
MOINDALVS
 
a brief introduction to Arima
Yue Xiangnan
 
Time series analysis 101
Balvinder Hira
 
Long Memory presentation to SURF
Richard Hunt
 
Different Models Used In Time Series - InsideAIML
VijaySharma802
 
Ad

Recently uploaded (20)

PDF
Driving Employee Engagement in a Hybrid World.pdf
Mia scott
 
PPTX
apidays Munich 2025 - Building an AWS Serverless Application with Terraform, ...
apidays
 
PPTX
apidays Helsinki & North 2025 - Running a Successful API Program: Best Practi...
apidays
 
PPTX
AI Presentation Tool Pitch Deck Presentation.pptx
ShyamPanthavoor1
 
PDF
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
apidays
 
PPTX
Aict presentation on dpplppp sjdhfh.pptx
vabaso5932
 
PPTX
apidays Singapore 2025 - Designing for Change, Julie Schiller (Google)
apidays
 
PPTX
apidays Singapore 2025 - From Data to Insights: Building AI-Powered Data APIs...
apidays
 
PDF
How to Connect Your On-Premises Site to AWS Using Site-to-Site VPN.pdf
Tamanna
 
PDF
OOPs with Java_unit2.pdf. sarthak bookkk
Sarthak964187
 
PDF
Product Management in HealthTech (Case Studies from SnappDoctor)
Hamed Shams
 
PPTX
Numbers of a nation: how we estimate population statistics | Accessible slides
Office for National Statistics
 
PDF
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
PPT
AI Future trends and opportunities_oct7v1.ppt
SHIKHAKMEHTA
 
PPTX
Listify-Intelligent-Voice-to-Catalog-Agent.pptx
nareshkottees
 
PPTX
Advanced_NLP_with_Transformers_PPT_final 50.pptx
Shiwani Gupta
 
PDF
Web Scraping with Google Gemini 2.0 .pdf
Tamanna
 
PPTX
apidays Helsinki & North 2025 - Vero APIs - Experiences of API development in...
apidays
 
PDF
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
PPTX
apidays Helsinki & North 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (A...
apidays
 
Driving Employee Engagement in a Hybrid World.pdf
Mia scott
 
apidays Munich 2025 - Building an AWS Serverless Application with Terraform, ...
apidays
 
apidays Helsinki & North 2025 - Running a Successful API Program: Best Practi...
apidays
 
AI Presentation Tool Pitch Deck Presentation.pptx
ShyamPanthavoor1
 
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
apidays
 
Aict presentation on dpplppp sjdhfh.pptx
vabaso5932
 
apidays Singapore 2025 - Designing for Change, Julie Schiller (Google)
apidays
 
apidays Singapore 2025 - From Data to Insights: Building AI-Powered Data APIs...
apidays
 
How to Connect Your On-Premises Site to AWS Using Site-to-Site VPN.pdf
Tamanna
 
OOPs with Java_unit2.pdf. sarthak bookkk
Sarthak964187
 
Product Management in HealthTech (Case Studies from SnappDoctor)
Hamed Shams
 
Numbers of a nation: how we estimate population statistics | Accessible slides
Office for National Statistics
 
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
AI Future trends and opportunities_oct7v1.ppt
SHIKHAKMEHTA
 
Listify-Intelligent-Voice-to-Catalog-Agent.pptx
nareshkottees
 
Advanced_NLP_with_Transformers_PPT_final 50.pptx
Shiwani Gupta
 
Web Scraping with Google Gemini 2.0 .pdf
Tamanna
 
apidays Helsinki & North 2025 - Vero APIs - Experiences of API development in...
apidays
 
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
apidays Helsinki & North 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (A...
apidays
 

EuroPython 2017 - PyData - Deep Learning your Broadband Network @ HOME

  • 2. Who am I? ● Machine Learning Engineer ○ Fraud Detection System ○ Software Defect Prediction ● Software Engineer ○ Email Services (40+ mil. users) ○ High traffic server (IPC, network, concurrent programming) ● MPhil, HKUST ○ Major : Software Engineering based on ML tech ○ Research interests : ML, NLP, IR
  • 3. Outline Data Collection Time series Analysis Forecast Modeling Anomaly Detection Naive approach Logging SpeedTest Data preparation Handling time series Seasonal Trend Decomposition Rolling Forecast Basic approaches Stationarity Autoregression, Moving Average Autocorrelation ARIMA Multivariate Gaussian LSTM
  • 7. Anomaly Detection (Naive approach in 2015)
  • 8. Problem definition ● Detect abnormal states of Home Network ● Anomaly detection for time series ○ Finding outlier data points relative to some usual signal
  • 9. Types of anomalies in time series ● Additive outliers
  • 10. Types of anomalies in time series ● Temporal changes
  • 11. Types of anomalies in time series ● Level shift
  • 12. Outline Data Collection Time series Analysis Forecast Modeling Anomaly Detection Naive approach Logging SpeedTest Data preparation Handling time series Seasonal Trend Decomposition Rolling Forecast Basic approaches Stationarity Autoregression, Moving Average Autocorrelation ARIMA Multivariate Gaussian LSTM
  • 13. Logging Data ● Speedtest-cli ● Every 5 minutes for 3 Month. ⇒ 20k observations. $ speedtest-cli --simple Ping: 35.811 ms Download: 68.08 Mbit/s Upload: 19.43 Mbit/s $ crontab -l */5 * * * * echo ‘>>> ‘$(date) >> $LOGFILE; speedtest-cli --simple >> $LOGFILE 2>&1
  • 14. Logging Data ● Log output $ more $LOGFILE >>> Thu Apr 13 10:35:01 KST 2017 Ping: 42.978 ms Download: 47.61 Mbit/s Upload: 18.97 Mbit/s >>> Thu Apr 13 10:40:01 KST 2017 Ping: 103.57 ms Download: 33.11 Mbit/s Upload: 18.95 Mbit/s >>> Thu Apr 13 10:45:01 KST 2017 Ping: 47.668 ms Download: 54.14 Mbit/s Upload: 4.01 Mbit/s
  • 15. Data preparation ● Parse data class SpeedTest(object): def __init__(self, string): self.__string = string self.__pos = 0 self.datetime = None# for DatetimeIndex self.ping = None # ping test in ms self.download = None# down speed in Mbit/sec self.upload = None # up speed in Mbit/sec def __iter__(self): return self def next(self): …
  • 16. Data preparation ● Build panda DataFrame speedtests = [st for st in SpeedTests(logstring)] dt_index = pd.date_range( speedtests[0].datetime.replace(second=0, microsecond=0), periods=len(speedtests), freq='5min') df = pd.DataFrame(index=dt_index, data=([st.ping, st.download, st.upload] for st in speedtests), columns=['ping','down','up'])
  • 18. Data preparation ● Structural breaks ○ Accidental missings for a long period
  • 19. Data preparation ● Handling missing data ○ Only a few occasional cases
  • 20. Handling time series ● By DatetimeIndex ○ df[‘2017-04’:’2017-06’] ○ df[‘2017-04’:] ○ df[‘2017-04-01 00:00:00’:] ○ df[df.index.weekday_name == ‘Monday’] ○ df[df.index.minute == 0] ● By TimeGrouper ○ df.groupby(pd.TimeGrouper(‘D’)) ○ df.groupby(pd.TimeGrouper(‘M’))
  • 21. Patterns in time series ● Is there a pattern in 24 hours?
  • 22. Patterns in time series ● Is there a daily pattern?
  • 23. Components of Time series data ● Trend :The increasing or decreasing direction in the series. ● Seasonality : The repeating in a period in the series. ● Noise : The random variation in the series.
  • 24. Components of Time series data ● A time series is a combination of these components. ○ yt = Tt + St + Nt (additive model) ○ yt = Tt × St × Nt (multiplicative model)
  • 25. Seasonal Trend Decomposition from statsmodels.tsa.seasonal import seasonal_decompose decomposition = seasonal_decompose(week_dn_ts) plt.plot(week_dn_ts) # Original plt.plot(decomposition.seasonal) plt.plot(decomposition.trend)
  • 27. Rolling Forecast from statsmodels.tsa.arima_model import ARIMA forecasts = list() history = [x for x in train_X] for t in range(len(test_X)): # for each new observation model = ARIMA(history, order=order) # update the model y_hat = model.fit().forecast() # forecast one step ahead forecasts.append(y_hat) # store predictions history.append(test_X[t]) # keep history updated
  • 28. Residuals ~ N( , 2 ) residuals = [test[t] - forecasts[t] for t in range(len(test_X))] residuals = pd.DataFrame(residuals) residuals.plot(kind=’kde’)
  • 29. Anomaly Detection (Basic approach) ● IQR (Inter Quartile Range) ● 2-5 Standard Deviation ● MAD (Median Absolute Deviation)
  • 30. Anomaly Detection (Naive approach) ● Inter Quartile Range
  • 31. Anomaly Detection (Naive approach) ● Inter Quartile Range ○ NumPy ○ Pandas q1, q3 = np.percentile(col, [25, 75]) iqr = q3 - q1 np.where((col < q1 - 1.5*iqr) | (col > q3 + 1.5*iqr)) q1 = df[‘col’].quantile(.25) q3 = df[‘col’].quantile(.75) iqr = q3 - q1 df.loc[~df[‘col’].between(q1-1.5*iqr, q3+1.5*iqr),’col’]
  • 32. Anomaly Detection (Naive approach) ● 2-5 Standard Deviation
  • 33. Anomaly Detection (Naive approach) ● 2-5 Standard Deviation ○ NumPy ○ Pandas std = pd[‘col’].std() med = pd[‘col’].median() df.loc[~df[‘col’].between(med - 3*std, med + 3*std), 0] std = np.std(col) med = np.median(col) np.where((col < med - 3*std) | (col < med + 3*std))
  • 34. Anomaly Detection (Naive approach) ● MAD (Median Absolute Deviation) ○ MAD = median(|Xi - median(X)|) ○ “Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median” - Christopher Leys (2013)
  • 35. Outline Data Collection Time series Analysis Forecast Modeling Anomaly Detection Naive approach Logging SpeedTest Data preparation Handling time series Seasonal Trend Decomposition Rolling Forecast Basic approaches Stationarity Autoregression, Moving Average Autocorrelation ARIMA Multivariate Gaussian LSTM
  • 36. Stationary Series Criterion ● The mean, variance and covariance of the series are time invariant. stationary non-stationary
  • 37. Stationary Series Criterion ● The mean, variance and covariance of the series are time invariant. stationary non-stationary
  • 38. Stationary Series Criterion ● The mean, variance and covariance of the series are time invariant. stationary non-stationary
  • 40. Differencing ● A non-stationary series can be made stationary after differencing. ● Instead of modelling the level, we model the change ● Instead of forecasting the level, we forecast the change ● I(d) = yt - yt-d ● AR + I + MA
  • 41. Autoregression (AR) ● Autoregression means developing a linear model that uses observations at previous time steps to predict observations at future time step. ● Because the regression model uses data from the same input variable at previous time steps, it is referred to as an autoregression
  • 42. Moving Average (MA) ● MA models look similar to the AR component, but it's dealing with different values. ● The model account for the possibility of a relationship between a variable and the residuals from previous periods.
  • 43. ARIMA(p, d, q) ● Autoregressive Integrated Moving Average ○ AR : A model that uses dependent relationship between an observation and some number of lagged observations. ○ I : The use of differencing of raw observations in order to make the time series stationary. ○ MA : A model that uses the dependency between an observation and a residual error from a MA model. ● parameters of ARIMA model ○ p : The number of lag observations included in the model ○ d : the degree of differencing, the number of times that raw observations are differenced ○ q : The size of moving average window.
  • 44. Identification of ARIMA ● Autocorrelation function(ACF) : measured by a simple correlation between current observation Yt and the observation p lags from the current one Yt-p . ● Partial Autocorrelation Function (PACF) : measured by the degree of association between Yt and Yt-p when the effects at other intermediate time lags between Yt and Yt-p are removed. ● Inference from ACF and PACF : theoretical ACFs and PACFs are available for various values of the lags of AR and MA components. Therefore, plotting ACFs and PACFs versus lags and comparing leads to the selection of the appropriate parameter p and q for ARIMA model
  • 45. Identification of ARIMA (easy case) ● General characteristics of theoretical ACFs and PACFs ● Reference : ○ http://people.duke.edu/~rnau/411arim3.htm ○ Prof. Robert Nau model ACF PACF AR(p) Tail off; Spikes decay towards zero Spikes cutoff to zero after lag p MA(q) Spikes cutoff to zero after lag q Tails off; Spikes decay towards zero ARMA(p,q) Tails off; Spikes decay towards zero Tails off; Spikes decay towards zero
  • 46. Identification of ARIMA (easy case)
  • 47. Identification of ARIMA (complicated)
  • 48. Anomaly Detection (Parameter Estimation) xdown xup xdown xup
  • 49. Anomaly Detection (Multivariate Gaussian Distribution)
  • 50. Anomaly Detection (Multivariate Gaussian) import numpy as np from scipy.stats import multivariate_normal def estimate_gaussian(dataset): mu = np.mean(dataset, axis=0) sigma = np.cov(dataset.T) return mu, sigma def multivariate_gaussian(dataset, mu, sigma): p = multivariate_normal(mean=mu, cov=sigma) return p.pdf(dataset) mu, sigma = estimate_gaussian(train_X) p = multivariate_gaussian(train_X, mu, sigma) anomalies = np.where(p < ep) # ep : threshold
  • 51. Anomaly Detection (Multivariate Gaussian) import numpy as np from scipy.stats import multivariate_normal def estimate_gaussian(dataset): mu = np.mean(dataset, axis=0) sigma = np.cov(dataset.T) return mu, sigma def multivariate_gaussian(dataset, mu, sigma): p = multivariate_normal(mean=mu, cov=sigma) return p.pdf(dataset) mu, sigma = estimate_gaussian(train_X) p = multivariate_gaussian(train_X, mu, sigma) anomalies = np.where(p < ep) # ep : threshold
  • 52. Anomaly Detection (Multivariate Gaussian) import numpy as np from scipy.stats import multivariate_normal def estimate_gaussian(dataset): mu = np.mean(dataset, axis=0) sigma = np.cov(dataset.T) return mu, sigma def multivariate_gaussian(dataset, mu, sigma): p = multivariate_normal(mean=mu, cov=sigma) return p.pdf(dataset) mu, sigma = estimate_gaussian(train_X) p = multivariate_gaussian(train_X, mu, sigma) anomalies = np.where(p < ep) # ep : threshold
  • 53. Outline Data Collection Time series Analysis Forecast Modeling Anomaly Detection Naive approach Logging SpeedTest Data preparation Handling time series Seasonal Trend Decomposition Rolling Forecast Basic approaches Stationarity Autoregression, Moving Average Autocorrelation ARIMA Multivariate Gaussian LSTM
  • 56. Long Short-Term Memory from keras.models import Sequential from keras.layers import Dense from keras.layers import LSTM from sklearn.metrics import mean_squared_error model = Sequential() model.add(LSTM(num_neurons, stateful=True, return_sequences=True, batch_input_shape=(batch_size, timesteps, input_dimension)) model.add(LSTM(num_neurons, stateful=True, batch_input_shape=(batch_size, timesteps, input_dimension)) model.add(Dense(1)) model.compile(loss='mean_squared_error', optimizer='adam') for i in range(num_epoch): model.fit(train_X, y, epochs=1, batch_size=batch_size, shuffle=False) model.reset_states()
  • 57. Long Short-Term Memory ● Will allow to model sophisticated and seasonal dependencies in time series ● Very helpful with multiple time series ● On going research, requires a lot of work to build model for time series
  • 58. Summary ● Be prepared before calling engineers for service failures ● Pythonista has all the powerful tools ○ pandas is great for handling time series ○ statsmodels for analyzing and modeling time series ○ sklearn is such a multi-tool in data science ○ keras is good to start deep learning ● Pythonista needs to understand a few concepts before using the tools ○ Stationarity in time series ○ Autoregressive and Moving Average ○ Means of forecasting, anomaly detection ● Deep Learning for forecasting time series ○ still on-going research ● Do try this at home