VAST Data Logo

VAST Data

Senior Solutions Engineer – Big Data & Data Infrastructure

Reposted 24 Days Ago
Be an Early Applicant
In-Office or Remote
Hiring Remotely in Tel Aviv
Mid level
In-Office or Remote
Hiring Remotely in Tel Aviv
Mid level
As a Senior Solutions Engineer, you'll design scalable data pipelines and cloud infrastructures, integrating real-time solutions and collaborating with teams to solve complex data challenges.
The summary above was generated by AI

This is a great opportunity to be part of one of the fastest-growing infrastructure companies in history, an organization that is in the center of the hurricane being created by the revolution in artificial intelligence.

"VAST's data management vision is the future of the market."- Forbes

VAST Data is the data platform company for the AI era. We are building the enterprise software infrastructure to capture, catalog, refine, enrich, and protect massive datasets and make them available for real-time data analysis and AI training and inference. Designed from the ground up to make AI simple to deploy and manage, VAST takes the cost and complexity out of deploying enterprise and AI infrastructure across data center, edge, and cloud.

Our success has been built through intense innovation, a customer-first mentality and a team of fearless VASTronauts who leverage their skills & experiences to make real market impact. This is an opportunity to be a key contributor at a pivotal time in our company’s growth and at a pivotal point in computing history.

We are seeking an experienced Solutions Data Engineer who possess both technical depth and strong interpersonal skills to partner with internal and external teams to develop scalable, flexible, and cutting-edge solutions. Solutions Engineers collaborate with operations and business development to help craft solutions to meet customer business problems.

A Solutions Engineer works to balance various aspects of the project, from safety to design. Additionally, a Solutions Engineer researches advanced technology regarding best practices in the field and seek to find cost-effective solutions.

Job Description:

We’re looking for a Solutions Engineer with deep experience in Big Data technologies, real-time data pipelines, and scalable infrastructure—someone who’s been delivering critical systems under pressure, and knows what it takes to bring complex data architectures to life. This isn’t just about checking boxes on tech stacks—it’s about solving real-world data problems, collaborating with smart people, and building robust, future-proof solutions.

In this role, you’ll partner closely with engineering, product, and customers to design and deliver high-impact systems that move, transform, and serve data at scale. You’ll help customers architect pipelines that are not only performant and cost-efficient but also easy to operate and evolve.

We want someone who’s comfortable switching hats between low-level debugging, high-level architecture, and communicating clearly with stakeholders of all technical levels.

Key Responsibilities:

  • Build distributed data pipelines using technologies like Kafka, Spark (batch & streaming), Python, Trino, Airflow, and S3-compatible data lakes—designed for scale, modularity, and seamless integration across real-time and batch workloads.
  • Design, deploy, and troubleshoot hybrid cloud/on-prem environments using Terraform, Docker, Kubernetes, and CI/CD automation tools.
  • Implement event-driven and serverless workflows with precise control over latency, throughput, and fault tolerance trade-offs.
  • Create technical guides, architecture docs, and demo pipelines to support onboarding, evangelize best practices, and accelerate adoption across engineering, product, and customer-facing teams.
  • Integrate data validation, observability tools, and governance directly into the pipeline lifecycle.
  • Own end-to-end platform lifecycle: ingestion → transformation → storage (Parquet/ORC on S3) → compute layer (Trino/Spark).
  • Benchmark and tune storage backends (S3/NFS/SMB) and compute layers for throughput, latency, and scalability using production datasets.
  • Work cross-functionally with R&D to push performance limits across interactive, streaming, and ML-ready analytics workloads.
  • Operate and debug object store–backed data lake infrastructure, enabling schema-on-read access, high-throughput ingestion, advanced searching strategies, and performance tuning for large-scale workloads.

Required Skills & Experience:

  • 2–4 years in software / solution or infrastructure engineering, with 2–4 years focused on building / maintaining large-scale data pipelines / storage & database solutions.
  • Proficiency in Trino, Spark (Structured Streaming & batch) and solid working knowledge of Apache Kafka.
  • Coding background in Python (must-have); familiarity with Bash and scripting tools is a plus.
  • Deep understanding of data storage architectures including SQL, NoSQL, and HDFS.
  • Solid grasp of DevOps practices, including containerization (Docker), orchestration (Kubernetes), and infrastructure provisioning (Terraform).
  • Experience with distributed systems, stream processing, and event-driven architecture.
  • Hands-on familiarity with benchmarking and performance profiling for storage systems, databases, and analytics engines.
  • Excellent communication skills—you’ll be expected to explain your thinking clearly, guide customer conversations, and collaborate across engineering and product teams.

Top Skills

Airflow
Docker
Kafka
Kubernetes
Python
S3
Spark
Terraform
Trino

VAST Data Campbell, California, USA Office

33 N 1st St, Campbell, California, United States, 95008

Similar Jobs

8 Hours Ago
Remote or Hybrid
Tel Aviv, ISR
Mid level
Mid level
Productivity • Sales • Software
Lead and build a development team for the CRM product, focusing on growth through effective collaboration and technical leadership. Responsibilities include managing projects, developing onboarding experiences, and iterating to optimize key business metrics.
Top Skills: Web Applications
2 Days Ago
Remote or Hybrid
Tel Aviv, ISR
Senior level
Senior level
Productivity • Sales • Software
The BI Developer will execute data insights, build DWH, manage ETLs, and collaborate with internal teams to drive growth.
Top Skills: AirflowBigQueryDatadogDbtDwhETLLookerLuigiRedshiftSnowflakeSQLTableau
3 Days Ago
Remote or Hybrid
Tel Aviv, ISR
Senior level
Senior level
Information Technology • Sales • Security • Cybersecurity • Automation
As FP&A Manager, you will coordinate business needs with financial goals, manage budget processes, and collaborate with departments to set strategic direction.
Top Skills: Bi SystemsExcelNetSuite

What you need to know about the San Francisco Tech Scene

San Francisco and the surrounding Bay Area attracts more startup funding than any other region in the world. Home to Stanford University and UC Berkeley, leading VC firms and several of the world’s most valuable companies, the Bay Area is the place to go for anyone looking to make it big in the tech industry. That said, San Francisco has a lot to offer beyond technology thanks to a thriving art and music scene, excellent food and a short drive to several of the country’s most beautiful recreational areas.

Key Facts About San Francisco Tech

  • Number of Tech Workers: 365,500; 13.9% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Google, Apple, Salesforce, Meta
  • Key Industries: Artificial intelligence, cloud computing, fintech, consumer technology, software
  • Funding Landscape: $50.5 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Sequoia Capital, Andreessen Horowitz, Bessemer Venture Partners, Greylock Partners, Khosla Ventures, Kleiner Perkins
  • Research Centers and Universities: Stanford University; University of California, Berkeley; University of San Francisco; Santa Clara University; Ames Research Center; Center for AI Safety; California Institute for Regenerative Medicine

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account