addressalign-toparrow-leftarrow-leftarrow-right-10x10arrow-rightbackbellblockcalendarcameraccwcheckchevron-downchevron-leftchevron-rightchevron-small-downchevron-small-leftchevron-small-rightchevron-small-upchevron-upcircle-with-checkcircle-with-crosscircle-with-pluscontroller-playcredit-cardcrossdots-three-verticaleditemptyheartexporteye-with-lineeyefacebookfolderfullheartglobe--smallglobegmailgooglegroupshelp-with-circleimageimagesinstagramFill 1languagelaunch-new-window--smalllight-bulblinklocation-pinlockm-swarmSearchmailmediummessagesminusmobilemoremuplabelShape 3 + Rectangle 1ShapeoutlookpersonJoin Group on CardStartprice-ribbonprintShapeShapeShapeShapeImported LayersImported LayersImported Layersshieldstartickettrashtriangle-downtriangle-uptwitteruserwarningyahooyoutube

MODUG: Mid-Ohio Data User Group Pages

August 25th Meetup
Presentations are available in the Files section

Looking at data thru Neo4j – David Fauth
Dave will provide a technical overview of Neo4j, discuss how RDMS and Neo4j differ, how to load data into Neo4j and how to query Neo4j using the Cypher Query Language.

Big Data Text Analytics – Victoria Lowengart
Text Analytics (TA) (which is a broader term for NLP, text mining, computational linguistics, information retrieval and information extraction) is the art and science of automatically or semi-automatically extracting meaningful information from unstructured text. With the exponential growth of unstructured text data (social media is one of the culprits here) Big Data technologies are imperative for the efficient and speedy text analytics tasks. In fact Big Data technologies infuse enhanced machine learning capabilities into Text Analytics

Smoothing the data rate peaks for Storm processing. - Chris Embree
When your Storm topology can process X bps and your peak ingest is >X, you need a fast, resilient input buffer to smooth things out and prevent data loss. Kafka allows you to do this with a small footprint and little care and feeding. We use Kafka to buffer 10Gbps data with millions of documents per second using a custom written producer. We also gather machine exhaust using Flume to write to Kafka topics. We’ll cover design and configuration topics.

Table of Contents

Page title Most recent update Last edited by
About MODUG: Mid-Ohio Data User Group August 28, 2015 10:53 AM Jeff G.

People in this
group are also in:

Sign up

Meetup members, Log in

By clicking "Sign up" or "Sign up using Facebook", you confirm that you accept our Terms of Service & Privacy Policy