We are a group of startup engineers, research scientists, computational linguists, mathematicians, philosophers, and others interested in understanding the meaning of text, reasoning, and human intent through technology. We want to apply our understanding to building new businesses and improving overall human experience in the modern connected world. The MIND Stack explained: mind.wtf.
This is a technical AI meetup: we build systems with Machine Learning on top of Data Pipelines, and concern ourselves with the stuff we can try in open source, learn, improve, and model human behavior in industry for practical results.
The advisory board for this meetup is Cicero Institute (Cicero.ai), and its conferences are AI.vision and self.driving.cars. We like specific technical problems (self-driving cars) and the way they inform better higher-level inference of the future of AI (AI.vision).
Machine Learning (ML) pipelines are the key building block for productionizing ML code. However, pipelines are often developed as "silos" - features tend not to be easily re-used across pipelines or even within the same pipeline. Silos lead to duplication, unnecessarily re-implementing features and in the worst case correctness problems, if, for example, the features used for training and serving have inconsistent implementations. The Feature Store solves the problem of siloed and ad-hoc machine learning pipelines, by providing a data layer where feature engineering can be separated from the usage of features to generate training data. That is, the Feature Store should provide a clean API separating Data Engineering from Data Science.
In this talk, we will introduce the world's first open-source Feature Store, built on Hopsworks, Apache Spark, and Apache Hive and targeting both TensorFlow/Keras and PyTorch. We will show how ML pipelines can be programmed, end-to-end, in Python, and the role of the Feature Store as a natural interface between Data Engineers and Data Scientists. In an end-to-end pipeline, we will show how the Feature Store works, and how you can write end-to-end ML pipelines in Python only (if you so choose).
Jim Dowling is the CEO of Logical Clocks AB, as well as an Associate Professor at KTH Royal Institute of Technology in Stockholm. He is the lead architect of Hops, the world's most fastest and most scalable Hadoop distribution and first Hadoop platform with support for GPUs as a resource. He is a regular speaker at AI industry conferences, and blogs at O'Reilly on AI.
Note: the limits of Deep Learning will be examined by the AI industry leaders at http://ai.vision conference, San Francisco, March 8. Early Bird ends 12/31.
This is a biweekly reading group for the book by Goodfellow, Bengio and Courville (https://www.amazon.com/Deep-Learning-Adaptive-Computation-Machine/dp/0262035618).
We'll meet in 2017 and go through it in sequence. We need a fixed venue to host us -- please email Alexy at [masked] if you want to be the home of the brave AI. We'll need reading group leaders for this to happen, please email Alexy if you want to lead certain chapters or the whole sequence.