Transform language models into real-world, high-impact product experiences.A1 is a self-funded AI group, operating in full stealth. We're building a new global consumer AI application focused on an important but underexplored use case.You will shape the core technical direction of A1 - model selection, training strategy, infrastructure, and long-term architecture. This is a founding technical role: your decisions will define our model stack, our data strategy, and our product capabilities for years ahead.You won't just fine-tune models - you'll design systems: training pipelines, evaluation frameworks, inference stacks, and scalable deployment architectures. You will have full autonomy to experiment with frontier models (LLaMA, Mistral, Qwen, Claude-compatible architectures) and build new approaches where existing ones fall short.Why This Role MattersYou are creating the intelligence layer of A1's first product, defining how it understands, reasons, and interacts with users.Your decisions shape our entire technical foundation — model architectures, training pipelines, inference systems, and long-term scalability.You will push beyond typical chatbot use cases, working on a problem space that requires original thinking, experimentation, and contrarian insight.You influence not just how the product works, but what it becomes, helping steer the direction of our earliest use cases.You are joining as a founding builder, setting engineering standards, contributing to culture, and helping create one of the most meaningful AI applications of this wave.What You'll DoBuild end-to-end training pipelines: data → training → eval → inferenceDesign new model architectures or adapt open-source frontier modelsFine-tune models using state-of-the-art methods (LoRA/QLoRA, SFT, DPO, distillation)Architect scalable inference systems using vLLM / TensorRT-LLM / DeepSpeedBuild data systems for high-quality synthetic and real-world training dataDevelop alignment, safety, and guardrail strategiesDesign evaluation frameworks across performance, robustness, safety, and biasOwn deployment: GPU optimization, latency reduction, scaling policiesShape early product direction, experiment with new use cases, and build AI-powered experiences from zeroExplore frontier techniques: retrieval-augmented training, mixture-of-experts, distillation, multi-agent orchestration, multimodal modelsWhat It's Like to Work HereYou take ownership - you solve problems end-to-end rather than wait for perfect instructionsYou learn through action - prototype → test → iterate → shipYou're calm in ambiguity - zero-to-one building energises youYou bias toward speed with discipline - V1 now > perfect laterYou see failures and feedback as essential to growthYou work with humility, curiosity, and a founder's mindsetYou lift the bar for yourself and your teammates every dayRequirementsStrong background in deep learning and transformer architecturesHands-on experience training or fine-tuning large models (LLMs or vision models)Proficiency with PyTorch, JAX, or TensorFlowExperience with distributed training frameworks (DeepSpeed, FSDP, Megatron, ZeRO, Ray)Strong software engineering skills — writing robust, production-grade systemsExperience with GPU optimization: memory efficiency, quantization, mixed precisionComfortable owning ambiguous, zero-to-one technical problems end-to-endNice to HaveExperience with LLM inference frameworks (vLLM, TensorRT-LLM, FasterTransformer)Contributions to open-source ML librariesBackground in scientific computing, compilers, or GPU kernelsExperience with RLHF pipelines (PPO, DPO, ORPO)Experience training or deploying multimodal or diffusion modelsExperience in large-scale data processing (Apache Arrow, Spark, Ray)Prior work in a research lab (Google Brain, DeepMind, FAIR, Anthropic, OpenAI)What You'll GetExtreme ownership and autonomy from day one - you define and build key model systems.Founding-level influence over technical direction, model architecture, and product strategy.Remote-first flexibilityHigh-impact scope—your work becomes core infrastructure of a global consumer AI product.Competitive compensation and performance-based bonusesBacking of a profitable US$2B group, with the speed of a startupInsurance coverage, flexible time off, and global travel insuranceOpportunity to shape a new global AI product from zeroA small, senior, high-performance team where you collaborate directly with founders and influence every major decision.Our Team & CultureWe operate as a dense, senior, high-performance team. We value clarity, speed, craftsmanship, and relentless ownership. We behave like founders — we build, ship, iterate, and hold ourselves to a high technical bar.If you value excellence, enjoy building real systems, and want to be part of a small team creating something globally impactful, you'll thrive here.About A1A1 is a self-funded, independent AI group backed by BJAK, focused on building a new consumer AI product with global impact. We're assembling a small, elite team of ML and engineering builders who want to work on meaningful, high-impact problems.