Skip to content

Project for the talk on NLP using LSTM implementation from DL4J on Spark

Notifications You must be signed in to change notification settings

kensuio-oss/NLP-LSTM-Spark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

spark-deep-learning

Deep Learning on Spark tutorial

Build

docker build -t bythebay/spark-deep-learning:0.0.1 .

Run

This uses a variable LOCAL_NOTEBOOKS which refers to a local directory containing the notebooks you want to include and keep up to date during the session.

Another folder you might want to sync is the data dir, which uses LOCAL_DATA then.

export LOCAL_NOTEBOOKS=<path to local notebooks dir>
export LOCAL_DATA=<path to local data dir>
docker run -v $LOCAL_NOTEBOOKS:/root/spark-notebook/notebooks/spark-deep-learning \
           -v $LOCAL_DATA:/root/data/spark-deep-learning \
           --rm -it -m 8g \
           -p 19000:9000 \
           -p 14040:4040 -p 14041:14041 -p 14042:4042 -p 14043:14043 \
           bythebay/spark-deep-learning:0.0.1 \
           bash

Then

source var.sh
source start.sh
source create.sh

Access

Open browser at http://localhost:19000/tree/spark-deep-learning.

WARN:

If you're running docker via a VM (Mac, Msft, ...) then you need to replace localhost by the VM's IP.

About

Project for the talk on NLP using LSTM implementation from DL4J on Spark

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published