Mount Indie is seeking a Data Engineer to support the design, development, and deployment of high-quality data pipelines and analytics for mission-focused applications.
This role integrates data engineering, exploratory data analysis, distributed systems, and cloud-native technologies to deliver performant, reliable, and secure data capabilities.
The Data Engineer will collaborate across multidisciplinary teams to ensure high-quality data output, robust testing, and operational readiness in support of DoD missions.
Responsibilities: * Conduct data pre-processing, exploratory data analysis, and data pipeline engineering to ensure performant and high-quality data output.
* Conduct thorough testing and validation of data pipelines and analytics to ensure accuracy, reliability, and robustness.
* Design or normalize data to common standards to support interoperability and analytical workflows.
* Develop and deploy data pipelines and analytics in real-world applications.
* Work with multiple data formats, including CSV, JSON, XML, Parquet, and ORC.
* Perform exploratory data analysis, algorithm development, and testing.
* Deploy, monitor, and improve data pipelines for operational environments.
* Implement event streaming pipelines using Apache Kafka, RabbitMQ, or ZeroMQ.
* Collaborate with analytics, engineering, and mission teams to ensure effective data integration and output quality.
* Stay current with emerging trends in data engineering, distributed systems, and modern data architecture.
* Document data processes, pipeline structures, and engineering best practices.
Required Skills and Experience: * At least 3 years of experience as a business analyst, data analyst, data scientist, data engineer, database administrator, geospatial analyst/engineer, machine learning engineer, or software engineer.
* Strong programming skills in Python.
* Experience designing or normalizing data to common standards.
* Experience with data pipeline development and real-world deployment.
* Experience with multiple data formats: CSV, JSON, XML, Parquet, ORC.
* Familiarity with event streaming platforms (Kafka, RabbitMQ, ZeroMQ). * Experience with exploratory data analysis, algorithm development, and testing.
* Experience deploying, monitoring, and improving data pipelines.
* Strong problem-solving and analytical skills.
* Excellent communication skills and ability to work effectively in a collaborative team environment.
* Familiarity with data pipeline frameworks and libraries (AirByte, Apache Airflow, dbt, Apache Iceberg, Snowflake). * Experience retrieving and managing GIS data (ArcGIS, PostGIS). * Programming skills in Go or Rust.
* Expertise with Elasticsearch, Redis, S3, PostgreSQL, or similar data stores.
* Experience with AWS native data services: EFS, RDS, S3, SNS, SQS.
* Experience with distributed computing and parallel processing (AWS Lambda, DASK, Spark). * Familiarity with cloud platforms (AWS, Azure) and containerization (Docker, Kubernetes). * Understanding of cybersecurity principles in the context of data applications.
* Previous experience with government agencies or military organizations.
Preferred Skills and Experience: * Experience with large-scale data architecture across secure DoD or government environments.
* Experience supporting NAVWAR, NIWC Pacific, or other Navy programs.
* Experience integrating data pipelines into operational mission systems.
* Familiarity with ML Ops or data engineering in classified or cross-domain environments.
Education and Certification Requirements: * Bachelor of Science in Computer Science, Data Science, Geography, Math, Machine Learning, or Statistics.
* Additional certifications in cloud, data engineering, GIS, or cybersecurity are a plus if required by contract.
Citizenship and Clearance Requirements: * US Citizenship * No dual citizenship * Active DoD TS/SCI clearance required Location: * Arlington, VA; Colorado Springs, CO; Charleston, SC; Denver, CO; or San Diego, CA.
Travel: * Up to 15% CONUS (and occasional OCONUS if mission requires).
Contractor