Let's face it - data sources are growing larger and more diverse every year - far outpacing the ability for us puny humans to successfully grok the data alone, let alone it's relevance to an investigation. One of the most important skills for an investigator or incident responder is that which will help him or her quickly answer diverse questions about the data they have been presented. Whether reviewing data from network captures, filesystem analysis, online services, or anywhere else - knowing how to slice data can give the investigator a clear edge in their pursuits. Most importantly, we can quickly eliminate the overwhelming volume of data that has no bearing on the investigation, leaving behind just the valuable tidbits that are most necessary. This talk will discuss how normalizing various data sources into a database can help wrangle data into a highly efficient tool that, when used properly, provides fast and decisive insight to the investigation at hand.
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
MOAR DATA\nMy HDDs\nMOAR DATA SOURCES\n- supertimelining = consistent picture from diverse data sources\n\n
Niche data sources too new or small to get attention from “big boys”\nScripting is possible, but even that can be limiting (and proprietary)\nFLEXIBLE!\n
Concept of foundational skill sets\nConstruction: Measuring, structural integrity\nCooking: Knife handling, food safety\nSales: Psychology, personality\nComputer Science: Data structures, algorithms\n
\n
SQL is an advanced art\nRaw power still accessible to anyone\n
\n
Data management\nDon’t re-invent if not needed\n“Can we fix it?”\n
Build for reduction\nSave space (and time!)\nKnow tradeoffs\n
Command Line Kung Fu\nUse what you know best!\n\n
Pitfalls with either choice\nBest of both: go big until too big\n
\n
Elegant/creative not always fast\nDOCUMENT!\n\n
\n
\n
\n
Only 800 records in VM\n
VM\n
VM\n
\n
Used schema with 8-10M sessions/day\n
150k records in VM\n
Observe to characterize traffic for reduction\n“always” reduce is bad practice (tunneling)\n
VM: Simple list of SSH sessions\nCould do Mbps\n
Two-step query\n(<200ms for both, ~10s for subSELECT)\n
Notional idea - untested\nALTER (can we fix it?)\nTrack running &#x201C;last seen&#x201D; date/time stamp\n