phData

Solutions Architect

phData

full-time

Posted on:

Origin:  • 🇮🇳 India

Visit company website
AI Apply
Manual Apply

Job Level

SeniorLead

Tech Stack

AirflowAWSAzureCassandraCloudElasticSearchGoogle Cloud PlatformHadoopHDFSInformaticaJavaKafkaMatillionNoSQLPythonScalaSparkSQL

About the role

  • Design and implement end-to-end cloud data solutions and pipelines for enterprise customers
  • Develop production-ready solutions ensuring performance, security, scalability, and robust data integration
  • Lead and mentor engineering teams and provide technical leadership
  • Serve as client-facing technical lead: create and deliver detailed presentations and communicate with stakeholders
  • Produce detailed solution documentation including POCs, roadmaps, sequence diagrams, class hierarchies, and logical system views
  • Collaborate with cloud and data platform teams using Snowflake, Spark, AWS, Azure, Databricks, and GCP
  • Support full software development lifecycle: design, implementation, testing, deployment

Requirements

  • 10+ years as a hands-on Solutions Architect and/or Data Engineer designing and implementing data solutions
  • Team lead, and/or mentorship of other engineers
  • Ability to develop end-to-end technical solutions into production — and to help ensure performance, security, scalability, and robust data integration
  • Programming expertise in Java, Python and/or Scala
  • Core cloud data platforms including Snowflake, Spark, AWS, Azure, Databricks and GCP
  • SQL and the ability to write, debug, and optimize SQL queries
  • Client-facing written and verbal communication skills and experience
  • Create and deliver detailed presentations
  • Detailed solution documentation (e.g. including POCS and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.)
  • 4-year Bachelor's degree in Computer Science or a related field
  • Preferred: Production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks
  • Preferred: Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra or other NoSQL storage systems
  • Preferred: Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc or other data integration technologies
  • Preferred: Multiple data sources (e.g. queues, relational databases, files, search, API)
  • Preferred: Complete software development lifecycle experience including design, documentation, implementation, testing, and deployment
  • Preferred: Automated data transformation and data curation: dbt, Spark, Spark streaming, automated pipelines
  • Preferred: Workflow Management and Orchestration: Airflow, AWS Managed Airflow, Luigi, NiFi