Data Solution Architect

US-VA-McLean

External

Req #: 6609
Type: Full-Time
logo

Steampunk

Connect With Us:
Connect To Our Company
				Overview:

We are looking for seasoned Data Solution Architect to work with our team and our clients to develop enterprise grade data platforms, services, pipelines, data models, visualizations, and more! The Data Solution Architect needs to be a technologist with excellent communication and customer service skills and a passion for data and problem solving. This role spans the spectrum of data capabilities, from data vision and strategy all the way through data science.

Responsibilities:

* Leading diverse technical teams comprised of data architects, data engineers, AI/ML engineers, data visualization specialists, and Cloud/DevOps engineers
* Designing greenfield data solution stacks in the cloud or on premises, using the latest data services, products, technology, and industry best practices
* Architecting migration of legacy data environments with performance and reliability
* Data Architecture contributions include assessing and understanding data sources, data models and schemas, and data workflows
* Data Engineering contributions include assessing, understanding, and designing ETL jobs, data pipelines, and workflows
* BI and Data Visualization contributions include assessing, understanding, and designing reports, selecting BI tools, creating dynamic dashboards, and setting up data pipelines in support of dashboards and reports
* Data Science contributions include assessing, understanding, and designing machine learning and AI applications, designing MLOps pipelines, and supporting data scientists
* Addressing technical inquiries concerning customization, integration, enterprise architecture and general feature / functionality of data products
* Experience in crafting data lakehouse solutions in the cloud (Preferably AWS. Alternatively, Azure, GCP). This includes relational databases, data warehouses, data lakes, and distributed data systems.
* Key must have skill sets - broad understanding of data exploitation lifecycle and capabilities
* Support an Agile software development lifecycle
* You will contribute to the growth of our AI & Data Exploitation Practice

Qualifications:

* Ability to hold a position of public trust with the US government.
* 12 years of directly related experience and a bachelor's degree OR 9 years and a masters degree
* Direct experience in Data Solutions with experience in tools such as:
* Big data tools: Hadoop, Spark, Kafka, etc.
* Relational SQL and NoSQL databases, such as Postgres, MySQL, MS SQL Server, Oracle, Mongo, etc.
* Data pipeline and workflow management tools: Airflow, NiFi, etc.
* AWS cloud services such as EC2, EMR, RDS, Redshift, Glue, SageMaker (or Azure and GCP equivalents)
* Data streaming systems: Storm, Spark-Streaming, etc.
* Data science tools/languages: R, R Studio, Python (data preparation and analysis libraries), Databricks, etc.
* Search tools: Solr, Lucene, Elasticsearch
* Object-oriented/scripting languages: Python, Java, C++, Scala, etc.

* Advanced working SQL knowledge and experience working with relational databases, query authoring and optimization (SQL) as well as working familiarity with a variety of databases.
* Advanced working Python and/or R
* Experience with DBOps and MLOps frameworks and exposure/understanding of DevOps
* Experience with message queuing, stream processing, etc
* Experience manipulating, processing, and extracting value from large, disconnected datasets.
* Experience manipulating structured and unstructured data for analysis
* Experience constructing complex queries to analyze results using databases or in a data processing development environment
* Experience with data modeling tools and process
* Experience architecting data systems (transactional and warehouses)
* Experience aggregating results and/or compiling information for reporting from multiple datasets
* Experience working in an Agile environment 
* Experience supporting project teams of developers and data scientists who build web-based interfaces, dashboards, reports, and analytics/machine learning models
			
Share this job: