
Explore more
Job Level
About the role
- Own and drive a structured, end-to-end CRO experimentation program aligned with business goals and user needs.
- Define conversion KPIs, success metrics, measurement plans, and reporting cadences.
- Partner closely with UX, product, and engineering teams to improve usability, reduce friction, and increase conversion efficiency across key journeys.
- Establish and maintain experimentation governance, including hypothesis creation, prioritization frameworks, QA processes, test documentation, and stakeholder reporting.
- Embed CRO best practices into information architecture, UX patterns, templates, and reusable components (e.g. CTA hierarchy, streamlined flows, friction reduction).****
- Define and implement robust measurement architecture, including funnel definitions, event schemas, attribution points, and baseline benchmarks.****
- Ensure experimentation readiness across platforms and tools, including:****
- - Stable DOM identifiers and test zones****
- - No-code / low-code testing compatibility****
- - Instrumentation QA and data integrity validation****
- Enable regionalized experimentation in alignment with localization and market-specific requirements.****
- Run post-launch optimization during the defined contract period by executing tests, analyzing results, and collaborating with engineering and content teams to roll out winning variants.****
- Maintain an ongoing experimentation cadence and document learnings, insights, and follow-on recommendations.****
- Executive-level CRO strategy and implementation documentation****
- CRO-ready templates and component standards (test zones and identifiers)****
- Funnel measurement plans and event specifications aligned with analytics and experimentation tooling****
- Experimentation governance toolkit (hypothesis templates, QA checklists, reporting formats)****
- Post-launch CRO operating plan and test cadence****
Requirements
- Strong experience designing and running structured CRO and A/B testing programs.****
- Hands-on expertise with experimentation platforms such as **VWO, Optimizely, Adobe Target**, and feature flagging tools (e.g. LaunchDarkly or similar).****
- Deep understanding of hypothesis-driven testing, test design standards, and statistical evaluation of results.
- Advanced knowledge of **GA4**, including event modeling, funnel analysis, and attribution.****
- Experience with data visualization and reporting tools such as **Looker Studio, Looker, or Tableau**.****
- Solid understanding of tagging and instrumentation using **GTM (client- and server-side)** and stable data attributes.****
- Ability to validate tracking accuracy, data quality, and experimentation integrity.****
- Strong UX and conversion optimization mindset, informed by qualitative tools such as **Hotjar, FullStory, Contentsquare, or Microsoft Clarity**.****
- Experience working with data warehouses and analysis tools such as **SQL, BigQuery, Snowflake**, or equivalent.****
- Proven ability to collaborate cross-functionally with UX/design, engineering, product, analytics, localization, and marketing automation teams.****
- Comfortable communicating insights and recommendations to both technical teams and executive-level stakeholders.****
Benefits
- **
- - Be part of an **AI-first, remote-first** digital agency that’s shaping the future of customer experiences.
- - Collaborate with global teams and leading platform partners to solve meaningful challenges.
- - Enjoy a culture that supports autonomy, continuous learning, and work-life harmony.
- **
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard Skills & Tools
CROA/B testinghypothesis-driven testingtest design standardsstatistical evaluationevent modelingfunnel analysisdata visualizationtagginginstrumentation
Soft Skills
collaborationcommunicationanalytical mindsetproblem-solvingcross-functional teamworkstakeholder engagementuser experience focusadaptabilitystrategic thinkingdocumentation