Why should you join dLocal?
dLocal enables the biggest companies in the world to collect payments in 40 countries in emerging markets. Global brands rely on us to increase conversion rates and simplify payment expansion effortlessly. As both a payments processor and a merchant of record where we operate, we make it possible for our merchants to make inroads into the world’s fastest-growing, emerging markets.
By joining us you will be a part of an amazing global team that makes it all happen. Being a part of dLocal means working with 1000+ teammates from 30+ different nationalities and developing an international career that impacts millions of people’s daily lives. We are builders, we never run from a challenge, we are customer-centric, and if this sounds like you, we know you will thrive in our team.
What's the Opportunity? You'll join the AI Lab, a team whose mission is to validate high-value emerging AI and automation technologies and de-risk their adoption across dLocal. This is a rare opportunity to work at the frontier of applied AI in fintech: running rigorous experiments on the latest models and tools, and turning results into decisions that shape how a global payments company like dLocal adopts AI.
As Technical Referent, you are the technical referent for technology scouting and evaluation within dLocal. You will run instrumented spikes and benchmarking on emerging AI technologies, produce clear recommendations for internal teams (could be business, legal, IT, etc.) at dLocal, and coordinate hand-offs to the teams that take validated technologies into production.
You will partner closely with engineering teams responsible for taking your proof of concepts into production (either domain teams or platformization teams): covering platform infrastructure, enablement programs, IT automations, and knowledge systems. Your core focus is evaluation and recommendation, not long-term ownership of production systems.
What Will I Be Doing?
Technology Scouting & Evaluation Run short, instrumented spikes and benchmarking on new models, tools and frameworks: LLMs, vector databases, orchestration frameworks, copilots, assistants and more. Compare vendor and open-source options, documenting trade-offs across quality, cost, latency, security and integration complexity. Deliver concise decision memos with clear recommendations: adopt, watch, or avoid. Evaluation Harnesses & Sandboxes Design and maintain evaluation environments (i.e. datasets, prompts, scenarios, telemetry) to test models under realistic constraints. Build automation and tooling to measure quality, robustness, latency and cost, including regression tracking over time. Ensure every evaluated technology has benchmark coverage and a documented risk and limitations view. Readiness Playbooks & Hand-offs For promising technologies, produce readiness playbooks describing recommended patterns, guardrails and integration guidelines. Coordinate with platformization teams to turn validated technologies into platformized capabilities. Track which validated items progress to platformization or pilots, and capture learnings to sharpen future bets. Governance, Risk & Standards Work with Security, Legal, Compliance and other AI teams to document risk assessments, mitigations and governance recommendations for each evaluated technology. Maintain checklists, decision templates and lightweight standards reusable across evaluations and by partner teams. Incorporate learnings from third-party AI tooling already in use (e.g., external copilots, AWS AI suite) into adoption guidelines. Collaboration, Mentoring & Community Partner with other AI teams and domain teams to ensure clear boundaries and smooth collaboration. Participate in hiring as a technical evaluator and culture champion. Mentor engineers in the Lab and adjacent teams on evaluation methods, benchmarking and experimental design. Share knowledge through internal write-ups, tech talks and occasional external meetups and conferences.
What Skills Do I Need?
Technical depth 8+ years in software engineering; 3+ years working hands-on with LLMs and AI tooling. Strong experience with distributed systems and event-driven architectures, both synchronous and asynchronous. Proficiency with LangChain, LangGraph or similar orchestration frameworks, including custom tools and multi-step workflows. Solid knowledge of AWS infrastructure and how to run evaluation workloads in a secure, cost-aware way. Track record designing and running benchmarks comparing AI models and tools under real constraints. Evaluation & decision-making Able to turn ambiguous "we should try this new thing" ideas into well-scoped evaluation plans with clear hypotheses and metrics. Comfortable making trade-off calls — quality vs. latency vs. cost vs. vendor lock-in — and documenting them clearly. Experience writing short, opinionated decision memos that help others move fast. Collaboration & communication Can explain technical results