Graduate Data Engineering Intern, Data Science

LinkedInMountain View, CA
401d$49 - $60Hybrid

About The Position

LinkedIn was built to help professionals achieve more in their careers, and everyday millions of people use our products to make connections, discover opportunities, and gain insights. Our global reach means we get to make a direct impact on the world's workforce in ways no other company can. We're much more than a digital resume - we transform lives through innovative products and technology. As a data engineer intern, you'll be transforming our data ecosystems. You will conduct a variety of applied research on the rich data that flows through our systems while effectively leveraging our data to create a single source of truth data. Successful candidates will exhibit technical acumen and business savvy, with a passion for making an impact through creative storytelling and timely actions. You will be working on our big data technology stack consisting of a variety of distributed platforms; we utilize both open-source and proprietary frameworks for large scale data processing including Hadoop, HDFS,Hive, and Spark. We also use Kafka for ingestion, Azkaban for workflow management, in addition to other applications. Candidates must be currently enrolled in a graduate degree program, with an expected graduation date of December 2025 or later. At LinkedIn, we trust each other to do our best work where it works best for us and our teams. This role offers a hybrid work option, meaning you can both work from home and commute to a LinkedIn office, depending on what's best for you and when it is important for your team to be together. Our internship roles will be based in Mountain View, CA; or other US office locations. Our internships are 12 weeks in length and will have the option of two intern sessions: * May 27th, 2025 - August 15th, 2025 * June 16th, 2025 - September 5th, 2025

Requirements

  • Currently pursuing a Graduate Degree in a quantitative discipline: computer science, statistics, applied mathematics, operations research, management of information systems, engineering, economics or equivalent and returning to the program after the completion of the internship.
  • Experience in at least one programming language (eg. Python, R, Hive, Java, Ruby, Scala/Spark or Perl etc.).
  • Experience with SQL or other relational databases.

Nice To Haves

  • Experience in Hadoop or other MapReduce paradigms and associated languages such as Pig and Hive.
  • Proven experience in developing data pipelines using Spark and Hive.
  • Experience with data modeling, ETL (Extraction, Transformation & Load) concepts, and patterns for efficient data governance.
  • Experience working with databases that power APIs for front-end applications.
  • Understanding data visualization tools (eg. Tableau, BI dashboarding, R visualization packages, etc.).
  • Experience building front-end visualizations using JavaScript frameworks (eg. jQuery, Marionette, D3, or Highcharts).
  • Experience in applied statistics and statistical modeling in at least one statistical software package, (eg. Advance R package, SAS, SPSS).
  • Ability to communicate findings clearly to both technical and non-technical audiences.

Responsibilities

  • Work with a team of high-performing data engineering professionals, and cross-functional teams to identify business opportunities and build scalable data solutions.
  • Build data expertise, act like an owner for the company and help manage complex data systems for a product or group of products.
  • Perform all of the necessary data transformations to serve products that empower data-driven decision making.
  • Establish efficient design and programming patterns for engineers as well as for non-technical partners.
  • Design, implement, integrate and document performant systems or components for data flows or applications that power analysis at a massive scale.
  • Understand the analytical objectives to make logical recommendations and drive informed actions.
  • Engage with internal data platform teams to prototype and validate tools developed in-house to derive insight from very large datasets or automate complex algorithms.

Benefits

  • The pay range for this role is $49 - $60 per hour.
  • The total compensation package for this position may also include annual performance bonus, stock, benefits and/or other applicable incentive compensation plans.

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Career Level

Intern

Industry

Administrative and Support Services

Education Level

Master's degree

Number of Employees

1-10 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service