Wynd Labs Logo

Wynd Labs

Data Engineer

Reposted 12 Days Ago
Remote
100K-140K
Mid level
Remote
100K-140K
Mid level
The Data Engineer will design and optimize data pipelines, integrate database systems, and automate workflows while collaborating with team members to ensure data accessibility and quality.
The summary above was generated by AI

Data Engineer

$100k - $140k

Who We Are.

Wynd Labs is an early-stage startup that is on a mission to make public web data accessible for AI through contributions to Grass.

Grass is a network sharing application that allows users to share their unused bandwidth. Effectively, this is a residential proxy network that directly rewards individual residential IPs for the bandwidth they provide. Grass will route traffic equitably among its network and meter the amount of data that each node provides to fairly distribute rewards.

In non-technical terms: Grass unlocks everyone's ability to earn rewards by simply sharing their unused internet bandwidth on personal devices (laptops, smartphones).

This project is for those who lead with initiative and seek to challenge themselves and thrive on curiosity.

We operate with a lean, highly motivated team who revel in the responsibility that comes with autonomy. We have a flat organizational structure, the people making decisions are also the ones implementing them. We are driven by ambitious goals and a strong sense of urgency. Leadership is given to those who show initiative, consistently deliver excellence and bring the best out of those around them. Join us if you want to set the tone for a fair and equitable internet.

The Role.

We are seeking a Data Engineer with expertise in building and maintaining robust data pipelines and integrating scalable infrastructure. You will join a small, talented team and play a critical role in designing and optimizing our data systems, ensuring seamless data flow and accessibility. Your contributions will directly support our mission to position Grass as a key player in the evolution of data-driven innovation on the internet.

Who You Are.

  • Bachelor’s degree in Computer Science, Information Systems, Data Engineering, or a related technical field.
  • Extensive experience with database systems such as Redshift, Snowflake, or similar cloud-based solutions.
  • Advanced proficiency in SQL and experience with optimizing complex queries for performance.
  • Hands-on experience with building and managing data pipelines using tools such as Apache Airflow, AWS Glue, or similar technologies.
  • Solid understanding of ETL (Extract, Transform, Load) processes and best practices for data integration.
  • Experience with infrastructure automation tools (e.g., Terraform, CloudFormation) for managing data ecosystems.
  • Knowledge of programming languages such as Python, Scala, or Java for pipeline orchestration and data manipulation.
  • Strong analytical and problem-solving skills, with an ability to troubleshoot and resolve data flow issues.
  • Familiarity with containerization (e.g., Docker) and orchestration (e.g., Kubernetes) technologies for data infrastructure deployment.
  • Collaborative team player with strong communication skills to work with cross-functional teams.

What You'll Be Doing.

  • Designing, building, and optimizing scalable data pipelines to process and integrate data from various sources in real-time or batch modes.
  • Developing and managing ETL/ELT workflows to transform raw data into structured formats for analysis and reporting.
  • Integrating and configuring database infrastructure, ensuring performance, scalability, and data security.
  • Automating data workflows and infrastructure setup using tools like Apache Airflow, Terraform, or similar.
  • Collaborating with data scientists, analysts, and other stakeholders to ensure efficient data accessibility and usability.
  • Monitoring, troubleshooting, and improving the performance of data pipelines and infrastructure to ensure data quality and flow consistency.
  • Working with cloud infrastructure (AWS, GCP, Azure) to manage databases, storage, and compute resources efficiently.
  • Implementing best practices for data governance, data security, and disaster recovery in all infrastructure designs.
  • Staying current with the latest trends and technologies in data engineering, pipeline automation, and infrastructure as code.

Why Work With Us.

  • Opportunity. We are at at the forefront of developing a web-scale crawler and knowledge graph that allows ordinary people to participate in the process, and share in the benefits of AI development.
  • Culture. We’re a lean team working together to achieve a very ambitious goal of improving access to public web data and distributing the value of AI to the people. We prioritize low ego and high output.
  • Compensation. You’ll receive a competitive salary and equity package.


Top Skills

Apache Airflow
AWS
Aws Glue
Azure
CloudFormation
Docker
GCP
Java
Kubernetes
Python
Redshift
Scala
Snowflake
SQL
Terraform

Similar Jobs

Yesterday
Remote
Hybrid
13 Locations
77K-129K Annually
Mid level
77K-129K Annually
Mid level
Artificial Intelligence • Healthtech • Machine Learning • Natural Language Processing • Biotech • Pharmaceutical
Responsible for data modeling, AI development, and optimizing data processes. Collaborate on digital outcomes and enhance data quality in healthcare solutions.
Top Skills: Amazon NeptuneAmazon RedshiftApache AirflowApache NifiApache SolrAWSAzureDockerElasticsearchGoogle BigqueryGoogle Cloud PlatformHadoopInformaticaJavaKafkaKubernetesNeo4JPrefectPythonScalaSnowflakeSparkSQLTalend
Yesterday
Remote
Hybrid
14 Locations
96K-161K Annually
Mid level
96K-161K Annually
Mid level
Artificial Intelligence • Healthtech • Machine Learning • Natural Language Processing • Biotech • Pharmaceutical
Manage data engineering projects focusing on AI and automation, develop AI models, optimize data pipelines, and lead data architecture efforts.
Top Skills: Amazon NeptuneAmazon RedshiftApache AirflowApache NifiApache SolrAWSAzureDockerElasticsearchGoogle BigqueryGoogle Cloud PlatformHadoopInformaticaJavaKafkaKubernetesNeo4JPrefectPythonScalaSnowflakeSparkSQLTalend
Yesterday
Remote
San Francisco, CA, USA
169K-271K Annually
Senior level
169K-271K Annually
Senior level
Cloud • Information Technology • Productivity • Security • Software • App development • Automation
Lead data engineering efforts at Atlassian, architect scalable data solutions, mentor engineers, and drive high-performing teams for analytics.
Top Skills: AirflowAWSDatabricksFlinkHiveKafkaRedshiftSparkSQL

What you need to know about the Boston Tech Scene

Boston is a powerhouse for technology innovation thanks to world-class research universities like MIT and Harvard and a robust pipeline of venture capital investment. Host to the first telephone call and one of the first general-purpose computers ever put into use, Boston is now a hub for biotechnology, robotics and artificial intelligence — though it’s also home to several B2B software giants. So it’s no surprise that the city consistently ranks among the greatest startup ecosystems in the world.

Key Facts About Boston Tech

  • Number of Tech Workers: 269,000; 9.4% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Thermo Fisher Scientific, Toast, Klaviyo, HubSpot, DraftKings
  • Key Industries: Artificial intelligence, biotechnology, robotics, software, aerospace
  • Funding Landscape: $15.7 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Summit Partners, Volition Capital, Bain Capital Ventures, MassVentures, Highland Capital Partners
  • Research Centers and Universities: MIT, Harvard University, Boston College, Tufts University, Boston University, Northeastern University, Smithsonian Astrophysical Observatory, National Bureau of Economic Research, Broad Institute, Lowell Center for Space Science & Technology, National Emerging Infectious Diseases Laboratories

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account