Do you want to tackle the biggest questions in finance with near infinite compute power at your fingertips?
G-Research is a leading quantitative research and technology firm, with offices in London and Dallas.
We are proud to employ some of the best people in their field and to nurture their talent in a dynamic, flexible and highly stimulating culture where world-beating ideas are cultivated and rewarded.
This is a role based in our new Soho Place office – opened in 2023 – in the heart of Central London and home to our Research Lab.
The role- Dates: 30 March 2026 – 18 September 2026 or 29 June 2026 – 18 September 2026
- Working hours: 09:00–17:30
- Location: Central London
We’re seeking a Software Engineering Intern to join our Core AI sub-team within the AI Engineering Group at G-Research.
The Core AI team builds and evolves the foundational platforms behind every Generative AI initiative within the firm, from RAG services to tooling that improves developer experience across quant and engineering teams.
As part of this team, you’ll contribute to the design and delivery of scalable, reliable and secure infrastructure that enables researchers, data scientists and engineers to experiment and deploy AI solutions safely and at speed.
Your work will span critical projects and workstreams, including distributed systems development, LLM orchestration and inference, RAG service integration and leading the drive for internal adoption of third-party AI technologies
Key responsibilities of the roll include:
- Designing, building and operating platform services in C# and Python, covering feature stores, vector search, prompt management and model hosting and inference
- Implementing type-safe orchestration workflows using LangGraph and Pydantic
- Scaling embedding and RAG pipelines to support large, high-throughput workloads
- Collaborating with research and product teams to productionise AI prototypes
- Applying best practices in testing, version control, CI/CD and observability
- Benchmarking and optimising latency, throughput and cost across on-prem GPU clusters and cloud platforms
We value engineers who are energised by complex, systems-level challenges, work fluently across languages and tooling and care deeply about developer experience and platform ergonomics.
You should be comfortable owning services end-to-end, from design docs to production dashboards, and motivated by the opportunity to help shape the foundations of AI development at G-Research.
The ideal candidate will have the following skills and experience:
- A degree in Computer Science, Engineering or a related field
- Strong production-grade programming skills in C# and Python, or equivalent
- Solid understanding of distributed systems, such as networking, storage, concurrency and fault tolerance
- Familiarity with AI engineering tooling, such as:
- LangGraph / LangChain
- Pydantic
- FastAPI
- MCP
- RAG pipelines
- Agentic workflows
The following experience is beneficial:
- Exposure to GPU scheduling or model-parallel inference (e.g. vLLM, TensorRT-LLM)
- Contributions to open-source AI infrastructure projects
- Highly competitive compensation plus annual discretionary bonus
- 30 days’ annual leave
- Lunch provided (via Just Eat for Business) and dedicated barista bar
- 9% company pension contributions
- Comprehensive healthcare and life assurance
- Informal dress code and excellent work/life balance
- Cycle-to-work scheme
- Monthly company events
G-Research is committed to cultivating and preserving an inclusive work environment. We are an ideas-driven business and we place great value on diversity of experience and opinions.
We want to ensure that applicants receive a recruitment experience that enables them to perform at their best. If you have a disability or special need that requires accommodation please let us know in the relevant section



