Diverse Lynx - Richardson, TX

posted 3 months ago

Full-time
Richardson, TX
Professional, Scientific, and Technical Services

About the position

The Python Developer role is a long-term contract position based in either Boston, MA or Richardson, TX. The ideal candidate will have a strong background in Python and SQL, with at least 5 years of experience in these areas. The role requires expertise in big data frameworks such as Hadoop and Spark, as well as experience in building automated data pipelines. Candidates should also have a solid understanding of data analysis and exploration techniques, and be comfortable working in an agile delivery environment. Strong critical thinking, communication, and problem-solving skills are essential for success in this position. In addition to the core skills, the candidate should have experience with Google Cloud Platform (GCP) and its services, including GCS Bucket, Cloud Functions, Dataproc, Dataflows, and Pub-Sub. Familiarity with version control systems like Git is necessary, as the role involves working in a multi-developer environment. The candidate should also have experience orchestrating data pipelines using tools such as Airflow or Azure Data Factory, and be knowledgeable about real-time and streaming technologies like Azure Event Hubs, Azure Functions, Kafka, and Spark Streaming. Furthermore, the role requires experience in API development and a basic understanding of DevOps best practices, including Continuous Integration and Continuous Deployment (CI/CD) using tools like Jenkins. Exposure to containerization technologies such as Kubernetes and Docker is also beneficial. This position offers an opportunity to work in a dynamic environment that values diversity and inclusion, and where all qualified applicants are considered based on their abilities and competencies.

Responsibilities

  • Develop and maintain Python applications for data processing and analysis.
  • Build automated data pipelines to streamline data workflows.
  • Perform data analysis and exploration to derive insights from large datasets.
  • Collaborate with team members in an agile delivery environment.
  • Utilize Google Cloud Platform services for data storage and processing.
  • Implement and manage data orchestration using tools like Airflow or Azure Data Factory.
  • Work with real-time and streaming technologies to handle data in motion.
  • Develop APIs to facilitate data access and integration.
  • Apply DevOps best practices to ensure efficient deployment and operation of applications.
  • Utilize version control systems to manage code changes in a multi-developer environment.

Requirements

  • 5+ years of experience in Python and SQL development.
  • Strong knowledge of big data frameworks such as Hadoop and Spark.
  • Experience in building automated data pipelines.
  • Proficient in data analysis and exploration techniques.
  • Experience working in an agile delivery environment.
  • Strong critical thinking, communication, and problem-solving skills.
  • Experience with Google Cloud Platform and its services.
  • Familiarity with version control systems like Git.
  • Experience with data orchestration tools such as Airflow or Azure Data Factory.
  • Knowledge of real-time and streaming technologies like Azure Event Hubs, Kafka, and Spark Streaming.
  • Experience in API development and understanding of DevOps best practices.

Nice-to-haves

  • Exposure to containerization technologies like Kubernetes and Docker.
  • Familiarity with CI/CD practices using Jenkins.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service