Insight Global is seeking a Remote Data Engineer to join one of their large government clients in Pennsylvania. The client is in the process of onboarding a new data system platform that will be used for research and analytics. This individual will be responsible for assisting with the expansion of the client's data pipeline architecture, as well as optimizing data flow and collection for the organization. The ideal candidate has atleast five years of experience in a Data Architect role and enjoys building and designing data systems from the ground up. This hire must be comfortable working alongside a team to ensure optimal data delivery architecture is consistent throughout the project. Additionally, they will help to lead conversations around system modernization and/or replacement efforts. This position is fully remote but will be working on east coast hours.
5+ years of experience in a Data Architect role
Strong data management skills to collect, store, and use data in an efficient and cost-effective manner
Experience in system development life cycle, project management approaches, and requirements, design, and test techniques
Experience in established and emerging data management and reporting technologies with knowledge of columnar and NoSQL databases, predictive analytics, data visualization, and unstructured data
Experience with big data tools: Hadoop, Spark, Kafka, etc.
Experience in artificial intelligence and machine learning (AI/ML) to build scalable systems for handling big data
Experience using data modeling tools like ERWin or Visio to visualize metadata and database schemas and/or architectures
Experience with relational SQL and NoSQL databases, including Oracle, MS SQL Server, Postgres, Cassandra, etc.
Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
Experience with data integration services solutions from vendors such as Informatica, MuleSoft, Talend, TIBCO, etc.
Experience with cloud-based data services such as AWS (EC2, Glue, EMR, RDS, Redshift, etc.) and/or Azure (Azure SQL, Data Factory, Synapse, Databricks, etc.)
Experience with stream-processing systems: Storm, Spark-Streaming, etc.
Experience with object-oriented/object function scripting languages: Python, R, Java, C++, Scala, etc.
Bachelor's Degree in Computer Science, Statistics, Informatics, Information Systems, or another quantitative field