ETL Developer

$85,800 - $209,100/Yr

CGI - Fairfax, VA

posted 3 months ago

Full-time - Mid Level
Hybrid - Fairfax, VA
1,001-5,000 employees
Professional, Scientific, and Technical Services

About the position

CGI Federal is seeking a motivated and customer-oriented ETL Developer to support a large federal client. This role involves designing, implementing, and managing ETL processes and data integration to generate valuable business insights. The position is based in our Fairfax, VA office, with a hybrid working model acceptable. In this role, you will be responsible for the design and development of ETL processes to extract data from various sources, transform it according to business requirements, and load it into target databases or data warehouses. You will develop and maintain ETL scripts, stored procedures, and functions, ensuring data integrity, consistency, and quality throughout the ETL process. You will also generate new business insights through data extraction, storage, transformation, analysis, and visualization of diverse data sets. This includes analyzing data using data mining, machine learning, and statistical algorithms available in commercial off-the-shelf (COTS) tools such as SAS and SPSS, as well as building analytical solutions using programming languages like R and Python. Your responsibilities will also include collecting and transforming structured, unstructured, relational, and NoSQL data using ETL and ELT tools, and developing custom code using programming languages. You will need to understand and use distributed methods such as MapReduce that scale to multi-Terabyte sized data collections. Additionally, you will apply data discovery and data visualization tools like Tableau and Trifacta to develop compelling, actionable, and useful data stories. Collaboration with data subject matter experts, business stakeholders, and management will be essential to prioritize business and information needs.

Responsibilities

  • Design and implement ETL processes to extract data from various sources, transform it according to business requirements, and load it into target databases or data warehouses.
  • Develop and maintain ETL scripts, stored procedures, and functions.
  • Generate new business insights through data extraction, storage, transformation, analysis, and visualization of diverse data sets.
  • Analyze data using data mining, machine learning, and statistical algorithms available in COTS tools (e.g., SAS, SPSS, Oracle).
  • Build analytical solutions using programming languages (e.g., R, Python, SAS) and programming libraries (e.g., Python SciKit, R Caret, PostgreSQL MADlib, Apache Spark MLlib).
  • Ensure data integrity, consistency, and quality throughout the ETL process.
  • Collect and transform structured, unstructured, relational, and NoSQL data using ETL and ELT tools, as well as develop custom code using programming languages.
  • Understand and use distributed methods (e.g., MapReduce) that scale to multi-Terabyte sized data collections.
  • Interpret and evaluate the accuracy of results through iterative, agile methods.
  • Apply data discovery and data visualization tools (e.g., Tableau, Trifacta) to develop compelling, actionable, and useful data stories.
  • Work closely with data SMEs, business, and management to prioritize business and information needs.

Requirements

  • 8+ years of relevant experience.
  • Experience working directly on a data science team.
  • Experience with tools, including object-oriented programming (Python, Java), computational analysis tools (R, MATLAB), and associated data science libraries (scikit-learn).
  • Proficiency in data science methods related to data architecture, data munging, data and feature engineering, and predictive analytics.
  • Experience creating meaningful data visualizations and interactive dashboards using platforms such as Tableau, Qlik, Power BI, RShiny, plotly, and d3.js.
  • Experience with cloud-based ETL solutions (e.g., AWS Glue, Azure Data Factory).
  • Knowledge of Big Data technologies (e.g., Hadoop, Spark).
  • Certification in relevant ETL or data integration technologies.
  • Bachelor's Degree in data science, mathematics, statistics, economics, computer science, engineering, or other related business or quantitative discipline.
  • Ability to obtain and maintain a Public Trust.

Nice-to-haves

  • Experience with additional data visualization tools beyond those listed in the requirements.
  • Familiarity with advanced machine learning techniques and frameworks.
  • Experience in a federal government contracting environment.

Benefits

  • 401(k) matching
  • Paid holidays
  • Paid parental leave
  • Paid time off
  • Tuition reimbursement
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service