
Description
Description
SAIC is seeking an experienced Senior Data Engineer to join our Innovation Lab Agile team delivering software solutions to the United States Patent and Trademark Office. In this role you will develop prototype Java applications of emerging technologies (AI/ML/NLP) that solve business problems. This resource will take ownership of building a modern composable data system using cutting-edge tools and technologies in the Apache Arrow ecosystem. As a Senior Data Engineer, you will play a critical role in designing, developing, and maintaining our data infrastructure. You will work closely with cross-functional teams to ensure our data systems are robust, scalable, and efficient. This role requires a deep understanding of big data systems and strong programming skills. join an energetic and versatile team responsible for addressing the customer challenges through the use of advancing technologies. This is a remote job that can be worked from any location in the continental U.S.
Key Responsibilities:
- Operate and maintain an innovation lab for the ideation, design, and engineering of cloud native product modernization and innovation using emerging technologies.
- Develop Java AWS based software proof of concept demonstrations that are compliant with USPTO cybersecurity policies.
- Develop and engineer pilots, prototypes, and proof of concept solutions to validate emerging technology, modernization, and innovation opportunities in an AWS environment.
- Design and implement a modern composable data stack using tools and technologies within the Apache Arrow ecosystem.
- Integrate and apply principles of composable data systems to enhance flexibility, scalability, and maintainability.
- Own and lead the implementation of data infrastructure projects.
- Collaborate with software engineers, data scientists, and other stakeholders to understand data needs and deliver solutions.
- Experience with Rust programming
- Ensure data systems are scalable, reliable, and maintainable.
- Write clean, efficient, and well-documented code.
- Optimize and tune data processes for performance and efficiency.
- Stay up-to-date with the latest industry trends and technologies in big data and data engineering
- Implement transparency, traceability, and appropriate security handling in line with best practices and applicable compliance standards including Authorization To Operate certifications.
- Collaborate with cross-functional teams including data engineering, product, and
- Support Agile product teams in emerging technologies and data strategies.
- Develop automated infrastructure as code and DevSecOps pipelines to streamline and automate the lab environment.
Qualifications
Required Qualifications:
- Bachelors' degree or higher in Computer Science, Data Science, Statistics, Mathematics, Engineering, or a related field.
- Minimum 7 years of relevant experience in data science, analytics, machine learning, artificial intelligence or a similar role.
- Candidates must be able to obtain and maintain a Public Trust clearance based on USPTO regulations
- Candidates must have lived in the United States for the last 2 years
- Strong programming skills in languages such as Java, C/C++, and Rust.
- Experience with Apache Arrow and related technologies.
- Familiarity with principles of composable data systems.
- Familiarity with data warehousing concepts and technologies.
- Excellent problem-solving skills and attention to detail.
- Strong communication and collaboration skills.
Technology Stack & Skills:
- Strong programming skills in languages such as Java, C/C++, and Rust.
- Experience with Apache Arrow and related technologies.
- Familiarity with principles of composable data systems.
- Familiarity with data warehousing concepts and technologies.
- Excellent problem-solving skills and attention to detail.
- Strong communication and collaboration skills.
- Data Manipulation & Analysis
- Visualization Tools Tableau, Power BI
- Big Data & Cloud Platforms: Hadoop, Hive, AWS (S3, SageMaker, Redshift), Azure, and GCP
- Version Control & Collaboration: Git, Docker
Desired Qualifications:
- Tasks may include developing data models for optimal performance in databases; designing data structures for data interchange; developing metadata structures including dictionaries, taxonomies and ontologies.
Target salary range: $120,001 - $160,000. The estimate displayed represents the typical salary range for this position based on experience and other factors.
Apply on company website