Similar documents
System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

arxiv: v1 [cs.lg] 7 Apr 2015

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

Georgetown University at TREC 2017 Dynamic Domain Track

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

arxiv: v4 [cs.cl] 28 Mar 2016

Residual Stacking of RNNs for Neural Machine Translation

Python Machine Learning

arxiv: v1 [cs.cl] 27 Apr 2016

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Dropout improves Recurrent Neural Networks for Handwriting Recognition

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH AND LANGUAGE PROCESSING, VOL XXX, NO. XXX,

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Second Exam: Natural Language Parsing with Neural Networks

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

A Case Study: News Classification Based on Term Frequency

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Improvements to the Pruning Behavior of DNN Acoustic Models

Australian Journal of Basic and Applied Sciences

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

arxiv: v5 [cs.ai] 18 Aug 2015

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

arxiv: v1 [cs.lg] 20 Mar 2017

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Generative models and adversarial training

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

arxiv: v1 [cs.cv] 10 May 2017

A Neural Network GUI Tested on Text-To-Phoneme Mapping

(Sub)Gradient Descent

Mathematics process categories

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Learning Methods for Fuzzy Systems

SORT: Second-Order Response Transform for Visual Recognition

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

NEURAL DIALOG STATE TRACKER FOR LARGE ONTOLOGIES BY ATTENTION MECHANISM. Youngsoo Jang*, Jiyeon Ham*, Byung-Jun Lee, Youngjae Chang, Kee-Eung Kim

Artificial Neural Networks written examination

ABC of Programming Linda

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Transfer Learning Action Models by Measuring the Similarity of Different Domains

arxiv: v1 [cs.lg] 15 Jun 2015

Using Deep Convolutional Neural Networks in Monte Carlo Tree Search

Knowledge Transfer in Deep Convolutional Neural Nets

Deep Neural Network Language Models

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Model Ensemble for Click Prediction in Bing Search Ads

The A2iA Multi-lingual Text Recognition System at the second Maurdor Evaluation

Reducing Features to Improve Bug Prediction

A Deep Bag-of-Features Model for Music Auto-Tagging

Beyond the Blend: Optimizing the Use of your Learning Technologies. Bryan Chapman, Chapman Alliance

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Evolutive Neural Net Fuzzy Filtering: Basic Description

Lecture 1: Machine Learning Basics

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

THE enormous growth of unstructured data, including

Assignment 1: Predicting Amazon Review Ratings

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Classification Using ANN: A Review

Lecture 1: Basic Concepts of Machine Learning

Calibration of Confidence Measures in Speech Recognition

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

On the Combined Behavior of Autonomous Resource Management Agents

SEMI-SUPERVISED ENSEMBLE DNN ACOUSTIC MODEL TRAINING

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Finding, Hiring, and Directing e-learning Voices Harlan Hogan, E-learningvoices.com

AI Agent for Ice Hockey Atari 2600

Major Milestones, Team Activities, and Individual Deliverables

Unit 3: Lesson 1 Decimals as Equal Divisions

Leveraging MOOCs to bring entrepreneurship and innovation to everyone on campus

ON THE USE OF WORD EMBEDDINGS ALONE TO

Cultivating DNN Diversity for Large Scale Video Labelling

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

CSL465/603 - Machine Learning

arxiv: v1 [cs.dc] 19 May 2017

arxiv: v2 [cs.cl] 26 Mar 2015

Institutionen för datavetenskap. Hardware test equipment utilization measurement

Forget catastrophic forgetting: AI that learns after deployment

Strategic Goals, Objectives, Strategies and Measures

An Introduction to Simio for Beginners

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Syllabus for CHEM 4660 Introduction to Computational Chemistry Spring 2010

Modeling function word errors in DNN-HMM based LVCSR systems

A Reinforcement Learning Variant for Control Scheduling

Getting Started with Deliberate Practice

Transcription:

1 Chart Pattern Matching in Financial Trading Using RNN Hitoshi Harada CTO hitoshi@alpacadb.com http://alpaca.ai Make you trade ideas into AI. Start free. On mobile. http://www.capitalico.com

What Technical Traders Are Looking For 2 Entry Point

Diversity Of The Pattern - All Downtrend 3

Problem And Needs - Fuzzy Pattern Recognition 4 Fuzzy Pattern Recognition for everyone Generalization (no hand crafted features) Multiple time series (OHLC price + indicators) Time scale, value scale, distortion Zhe Zhang, Jian Jiang, Xiaoyan Liu, Ricky Lau, Huaiqing Wang, Rui Zhang: A Real Time Hybrid Pattern Matching Scheme for Stock Time Series, 2010 James N.K. Liu *, Raymond W.M. Kwong : Automatic extraction and identification of chart patterns towards financial forecast, 2006

How To Solve The Problem? 5 ah p down trend SPEECH RECOGNITION WITH DEEP RECURRENT NEURAL NETWORKS, Hinton, et al. 2013 Capitalico

Interactive Training Data Collection & Training 6

Our Approach - Fuzzy Pattern Recognition without Programming 7 Train by what you see & judge No programming nor conditional setting, but purely from charts like traders do Multi-dimensional input Not only the single timeseries data of price movement but also various indicators altogether

Experiments Deep Learning Based Approach 8 Network Output Input: N-dim Fully Connected Layer Output Sigmoid LSTM Layer x 2 or 4 ( x250 units ) Fully Connected Layer ( x250 units ) Dropout Output Sigmoid Fully Connected Sigmoid Output: 1-dim confidence level Sigmoid Fully Connected LSTM LSTM Training Align with fixed number of candles Mean squared error for loss AdaDelta for optimizer Fully Connected LSTM LSTM LSTM LSTM Fully Connected Fully Connected Input BPTT through aligned length Fully Connected Input Data 1k+ samples collected by experts Input about hundred instances for each strategy Time

y-axis: confidence E xpe r i m e nt s Fit t ing Reas o n abl y x-axis: time (1.0=entry point) blue: training data / orange: testing data 9

Experiments Framework 10

Dropout 11 Dropout vs # of training samples Bigger Mini-Batches by looping samples Made it Adaptive depending on importance dropout enabled (x: iteration count, y: loss) dropout w/ bigger mini-batches (x: iteration count, y: loss)

Forget Gate Bias (Learning To Forget: Continual Prediction With Lstm, Felix Et Al.) 12

Trial And Error To Speed Up Training 13 Dynamic Dropout Dynamic Batchsize Multi-GPU Training Other Frameworks like Keras GRU IRNN Lot more

Conclusion & Future Work 14 14 Previous studies have limitations to difficulty of feature crafting. LSTM based deep neural network fits well with individual patterns. LSTM-variant doesn t make much difference, but forget-gate bias, normalization, preprocessing, and modeling etc. matter Build better base model by pre-training Reinforcement Learning using profit and risk preference Visualize and rationalize LSTM decision making Generative Model

QUESTIONS AND ANSWERS Make you trade ideas into AI. Start free. On mobile. http://www.capitalico.com http://alpaca.ai / info@alpacadb.com

References 16 Ken-ichi Kainijo and Tetsuji Tanigawa: Stock Price Pattern Recognition - A Recurrent Neural Network Approach -, 1990 S Hochreiter, J Schmidhuber: Long short-term memory, 1997 FA Gers, J Schmidhuber, F Cummins: Learning to forget: Continual prediction with LSTM, 2000 James N.K. Liu *, Raymond W.M. Kwong: Automatic extraction and identification of chart patterns towards financial forecast, 2006 X Guo, X Liang, X Li: A stock pattern recognition algorithm based on neural networks, 2007 Z Zhang, J Jiang, X Liu, R Lau, H Wang: A real time hybrid pattern matching scheme for stock time series, 2010 A Graves, A Mohamed, G Hinton: Speech recognition with deep recurrent neural networks, 2013 A Graves, N Jaitly: A Mohamed, Hybrid speech recognition with deep bidirectional LSTM, 2013 Tara N. Sainath, Oriol Vinyals, Andrew Senior, Has im Sak: CONVOLUTIONAL, LONG SHORT-TERM MEMORY, FULLY CONNECTED DEEP NEURAL NETWORKS

Need For Gpu And Distributed Computation 17 Model Training Takes around 10 minutes on a single GPU core Requires 2GB of GPU RAM Backtesting Calculate various metrics over the historical data Livetesting Thousands of models need to monitor live candles and update the state of LSTM

Need For Distributed Computation 18 DB Postgresql Redis etcd Market Data Historical Real time Load Balancer WEB Flask Live Queue Celery WORKER Algos = Market Watch ~10MB tesla k80 x1-10k Trading