28 June 2016

Big Data Project with Microservices (and Task) Description

This description was collected and forwarded by Artech, a staffing firm.
  1.        Build the Event Hubs integration with Service Fabric microservices implementation. Streaming the processed files from blobs into EH for downstream processing.
    1.        Anonymized files (~1000 of them and to a size of ~GB) will be given as input
    2.       Service Fabric code portion will be provided.
  2.        Build the Spark processing reading off EventHubs, implementation in either Python or Scala would suffice.
    1.        Look at the caching needs; leverage .cache to retain appropriate results from Spark ‘Actions’ in Spark executors
  3.        Our team will evaluate a set of data store that would be a landing spot post Spark – Blobs being a required one. We will pick 1 or 2 from this list -- SQL DW, Azure SQL DB, Cassandra and DocumentDB being other candidate stores and we will have code snippets and/or guidance
 Integration & Deployment
  1.        Integrate the items from above with completed items (Azure Data Factory with ARM provisioning, picking up from the ADF pipeline which lands files onto blobs)
  2.        Apply best practices for capacity planning, deployment for E2E
  3.        Integrate the deployment with existing set of tools and processes.  
  1.        Build a unit test framework that can test each building block in isolation (ADF à Blobs, Blobs à Service Fabric, Service Fabric à EH, EH à Spark, Spark à <Data Store>
  2.        Build an E2E test environment with telemetry on latency, throughout with percentiles. *Leverage APM tools as appropriate 

16 June 2016

Cross-Application SIEM Position is Well-Focused

logo for TRIGYN on KnowledgeGigs.com
The following post by Trigyn appeared on Monster. The job title as advertised: "Security Analyst (SIEM / Security Log-specific Design) at Trigyn Technologies Inc."


• Role is to engage approximately 35 high priority App/Technology owners, and:• Identify log generation and collection build requirements for their application/service-specific logs• Identify log generation and collection build requirements for any of their technology/infrastructure-specific logs that have somehow been missed in the current infrastructure log collection activity• Identify data fields & example values for the content of their logs (log-specific data dictionary)• Provide application on-boarding prioritization recommendations, based upon:• Available capacity in SIEM environment vs. log capacity/build requirements for their application(s)• Timeline to log availability• Identify Security Monitoring detection/alerting use cases of specific value to at least 4 Tier I Applications.