Hydraforce - Saratoga, CA

posted about 1 month ago

Full-time - Senior
Saratoga, CA
10,001+ employees
Fabricated Metal Product Manufacturing

About the position

The Data Architect position at Bosch Global Software Technologies Private Limited involves designing and implementing high-performance data architectures to support generative AI solutions. The role requires collaboration with Generative AI Engineers to create scalable and efficient data infrastructures, focusing on cutting-edge technologies such as vector databases and infrastructure as code. The Data Architect will play a crucial role in shaping the future of AI by ensuring data quality, governance, and performance optimization.

Responsibilities

  • Partner with Generative AI Engineers and architects to understand their data requirements and design a highly scalable, secure, and efficient data architecture.
  • Design and implement data pipelines for ingesting, processing, and storing massive datasets for training and running generative models.
  • Select and implement cutting-edge data storage solutions, considering scalability, performance, cost, and suitability for vector data.
  • Ensure data quality by implementing data cleansing, transformation, and validation processes.
  • Develop data governance policies and procedures to ensure data security, compliance, and accessibility.
  • Automate data pipelines and workflows using tools optimized for high-performance data processing.
  • Monitor and optimize data infrastructure performance for efficiency and scalability.
  • Collaborate with Data Scientists and Machine Learning Engineers to understand broader data needs and ensure alignment.
  • Stay up-to-date on the latest big data technologies, vector databases, and best practices for data management in AI environments.

Requirements

  • 10 years of experience in data architecture design and implementation, with a focus on high-performance data solutions
  • Strong understanding of data management principles, data modeling techniques, data governance practices, and distributed systems
  • Experience working with big data technologies (e.g., Kafka, PostgreSQL, MongoDB) and familiarity with vector databases (e.g., Pinecone, Faiss, Lance DB)
  • Proficiency in SQL and experience with data warehousing solutions (e.g., Snowflake, Redshift) is an added advantage
  • Experience with Azure and AWS cloud platforms and Terraform
  • Excellent communication and collaboration skills to effectively interact with technical and non-technical stakeholders
  • Strong problem-solving and analytical skills with a data-driven approach
  • Ability to work independently and manage multiple projects simultaneously
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service