Senior Data Engineer – Databricks | Remote-Friendly
At Velotio, we are embracing a remote-friendly work culture where everyone has the flexibility to either work remotely or from our office in Pune.
Join us and work from wherever you feel most productive!
About Velotio:
Velotio Technologies is a product engineering company working with innovative startups and enterprises. We are a certified Great Place to Work® and recognized as one of the best companies to work for in India. We have provided full-stack product development for 110+ startups across the globe building products in the cloud-native, data engineering, B2B SaaS, IoT & Machine Learning space. Our team of 325+ elite software engineers solves hard technical problems while transforming customer ideas into successful products.
Requirements
- Work closely with our clients providing evaluation and recommendations of design patterns and solutions for data platforms with a focus on ETL, ELT, ALT, lambda, and kappa architectures
- Define SLAs, SLIs, and SLOs with inputs from clients, product owners, and engineers to deliver data-driven interactive experiences
- Provide expertise, proof-of-concept, prototype, and reference implementations of architectural solutions for cloud, on-prem, hybrid, and edge-based data platforms
- Provide technical inputs to agile processes, such as epic, story, and task definition to resolve issues and remove barriers throughout the lifecycle of client engagements
- Creation and maintenance of infrastructure-as-code for cloud, on-prem, and hybrid environments using tools such as Terraform, CloudFormation, Azure Resource Manager, Helm, and Google Cloud Deployment Manager
- Mentor, support and manage team members.
Desired Skills & Experience:
- 6+ years of demonstrable experience in enterprise level data platforms involving implementation of end-to-end data pipelines
- Hands-on experience in using Databricks
- Hands-on experience with at least one of the leading public cloud data platforms (Amazon Web Services, Azure or Google Cloud)
- Experience with column-oriented database technologies (e.g., Big Query, Redshift, Vertica), NoSQL database technologies (e.g., DynamoDB, BigTable, Cosmos DB, etc.) and traditional database systems (e.g., SQL Server, Oracle, MySQL)
- Experience in architecting data pipelines and solutions for both streaming and batch integrations using tools/frameworks like Glue ETL, Lambda, Google Cloud DataFlow, Azure Data Factory, Spark, Spark Streaming, etc.
- Metadata definition and management via data catalogs, service catalogs, and stewardship tools such as OpenMetadata, DataHub, Alation, AWS Glue Catalog, Google Data Catalog.
- Test plan creation and test programming using automated testing frameworks, data validation and quality frameworks, and data lineage frameworks
- Data modeling, querying, and optimization for relational, NoSQL, timeseries, graph databases, data warehouses and data lakes
- Data processing programming using SQL, DBT, Python, and similar tools
- Logical understanding of programming in Python, Spark, PySpark, Java, Javascript, and/or Scala
- Cloud-native data platform design with a focus on streaming and event-driven architectures
- Participate in integrated validation and analysis sessions of components and subsystems on production servers
- Data ingest, validation, and enrichment pipeline design and implementation
- SDLC optimization across workstreams within a solution
- Bachelor’s degree in Computer Science, Engineering, or related field
Benefits
Our Culture:
- We have an autonomous and empowered work culture encouraging individuals to take ownership and grow quickly
- Flat hierarchy with fast decision making and a startup-oriented “get things done” culture
- A strong, fun & positive environment with regular celebrations of our success. We pride ourselves in creating an inclusive, diverse & authentic environment
Note: Currently, all interviews and onboarding processes at Velotio are being carried out remotely through virtual meetings.