Current Statistics
1,607,645 Total Jobs 333,747 Jobs Today 16,998 Cities 222,734 Job Seekers 146,858 Resumes |
|
|
![](/images/main_topleft.gif) |
|
![](/images/main_topright.gif) |
![](/images/main_left.jpg) |
Platform & HPC Data Engineer (TS/SCI with CI Poly) - Herndon Virginia
Company: Maxar Technologies Location: Herndon, Virginia
Posted On: 01/24/2025
Please review the job details below. Maxar is seeking a skilled Platform and HPC Data Engineer --to support the design, implementation, and optimization of data management solutions in high-performance computing (HPC) environments. The ideal candidate will have extensive experience working with various file systems, data labeling/tagging systems, and the configuration of a wide range of storage appliances. This role involves ensuring that data workflows, storage configurations, and metadata management are efficient, scalable, and aligned with organizational and government security requirements. The successful candidate will work within a cross-disciplinary team to support the technical needs of HPC platforms, data management, and large-scale computational workflows. Key Responsibilities: - Platform and HPC Data Engineering: --Design and implement data management systems and architectures for HPC platforms, focusing on optimizing data flow, storage, and access in large-scale computing environments.
- File System Management: Oversee the configuration, maintenance, and optimization of distributed file systems (e.g., Lustre, IBM Spectrum Scale, NFS, GPFS) and storage solutions used in HPC environments to ensure efficient performance, scalability, and reliability.
- Data Labeling and Tagging: --Implement and manage metadata-driven systems for data labeling/tagging. This includes the development of strategies for classifying, indexing, and organizing datasets to enhance data discoverability, access control, and auditing.
- Storage Appliance Configuration: Configure and maintain various storage appliances (e.g., NetApp, Dell EMC, HPE) and integrated storage solutions. Ensure that storage devices are optimized for performance, capacity, and availability within the HPC ecosystem.
- Data Integration and Workflow Optimization: Integrate data storage and management systems with HPC clusters, ensuring seamless data flow between compute nodes and storage appliances. Optimize data pipelines to support high-throughput workloads and minimize bottlenecks in I/O performance.
- Performance Tuning: --Monitor and improve the performance of storage systems, focusing on I/O throughput, latency, and efficient resource allocation. Use performance metrics to guide optimizations across storage appliances and file systems.
- Security and Compliance: --Implement security best practices for data access, protection, and management, ensuring compliance with government regulations and internal data governance policies. Configure encryption, access control, and secure data sharing methods.
- A utomation and Scripting: Develop and maintain automation scripts (e.g., using Python, Bash, or Perl) to streamline storage configurations, data labeling/tagging, and system monitoring tasks. Automate processes related to data integration and HPC platform management.
- Collaboration and Support: Work closely with data scientists, HPC administrators, software developers, and other technical staff to support ongoing projects. Provide expertise in troubleshooting data storage issues and ensuring optimal system performance.
- Documentation and Reporting: --Maintain thorough documentation for storage configurations, file system setups, data labeling/tagging procedures, and performance optimization strategies. Provide regular reports on system health, data management processes, and any improvements made. Required Qualifications:
- Education:--Bachelor's degree in Computer Science, Information Technology, Engineering, or a related field. A Master's degree or higher is a plus.
- Experience:
- 7+ years of experience in managing data infrastructure in HPC environments, with expertise in file systems, storage appliances, and data workflows.
- Hands-on experience with distributed file systems, including Lustre, IBM Spectrum Scale (GPFS), NFS, and others commonly used in HPC settings.
- Proven experience with storage appliance configuration (e.g., NetApp, Dell EMC, HPE, or similar systems), including performance tuning, capacity management, and reliability.
- Strong experience in implementing data labeling/tagging systems, metadata management, and structuring large datasets for efficient access and compliance.
- Knowledge of high-performance networking protocols (e.g., InfiniBand, RDMA) and their role in data transfer and storage optimization.
- Familiarity with data access protocols like GridFTP, rsync, and NFS for large-scale data transfer. Desired Skills:
|
![](/images/main_right.jpg) |
![](/images/main_botleft.gif) |
![](/images/main_bot.gif) |
![](/images/main_botright.gif) |
|
|