Tech Stack
AirflowApacheAWSAzureCloudDockerETLGoogle Cloud PlatformGRPCJavaKafkaKubernetesMicroservicesNoSQLPythonPyTorchSparkSQLTensorflow
About the role
- Engage with stakeholders to understand product requirements and architect endtoend solutions incorporating data ingestion, model training, deployment and monitoring.
- Design systems that combine traditional software components with AI/ML models and modern data engineering practices.
- Write and maintain production-quality code across backend services, data pipelines and model inference layers, primarily using Python and at least one other language.
- Build RESTful or gRPC APIs to expose AI functionality and integrate with the frontend and other microservices.
- Develop and manage data pipelines for ingestion, transformation and storage (using tools such as Apache Spark, Kafka, Airflow, SQL/NoSQL databases).
- Work with vector databases and embedding stores for retrieval-augmented systems; implement embedding techniques and RAG pipelines.
- Train, finetune and deploy machinelearning and deeplearning models, including large language models (LLMs) and transformer architectures (e.g., GPT, Llama, Mistral).
- Optimise model performance (latency, throughput and accuracy) using techniques such as model quantisation, batching and caching; scale services using containerisation (Docker) and orchestration (Kubernetes) on cloud platforms (AWS, Azure, GCP).
- Collaborate with software engineers, data scientists and DevOps teams to deliver cohesive solutions; mentor junior engineers and share best practices.
- Implement observability for data pipelines and AI services (logging, metrics, model drift detection) and continuously monitor and iterate models and pipelines based on feedback and new data.
Requirements
- Bachelors or Masters degree in Computer Science, Data Engineering, Artificial Intelligence or related field.
- Minimum 6 to 8 years in software development and data engineering, with at least 3 years of handson AI/ML experience.
- Experience : 5 + years
- Proficiency in Python and at least one other language (e.g., .Net, Java).
- Deep familiarity with AI/ML frameworks (PyTorch, TensorFlow, Scikitlearn) and LLM libraries (Hugging Face Transformers, LangChain).
- Handson experience building and deploying large language models, embeddings, RAG pipelines and vectordatabasepowered applications.
- Strong knowledge of data engineering tools (Spark, Kafka, Airflow) and relational and NoSQL databases; proficiency in SQL.
- Experience designing and operating microservices, APIs and containerised applications on cloud platforms (AWS, Azure, GCP).
- Solid understanding of MLOps best practices (CI/CD for ML, model versioning, automated testing, monitoring).
- Strong problemsolving and communication skills; ability to translate business requirements into technical designs.
- Work Mode: Onsite Karachi; Work Timings: Monday to Fri: 11:00 AM - 7:30 PM.