Who We Are
Bright Fox AI is a team of cloud architects, machine-learning theorists, and energy veterans turning the latest research on large language models (LLMs) into production-grade digital employees.
About the Role
You’ll be an engineer operating where LLM research meets cloud-native scale. Your charter is to design and own the infrastructure that trains, fine-tunes, and serves massive models—then squeeze every watt of insight out of them for our energy clients.
If you can diagram a Transformer’s attention flow, debate LoRA vs. QLoRA compression, and spin up a GPU fleet with in cloud environments, come build the future with us.
In this role you will tackle in these areas:
Cloud & DevOps
• Architect multi-cloud (AWS / Azure / GCP) GPU platforms • Embed GitOps pipelines (ArgoCD, GitHub Actions) for end-to-end model lifecycle
LLM Systems
• Fine-tune and distill GPT-class models using DeepSpeed, FSDP, and PEFT (LoRA/QLoRA)• Build Retrieval-Augmented Generation (RAG) pipelines with vector DBs (FAISS, Weaviate, …)• Construct LLM applications with frameworks such as LangChain, LlamaIndex and multi-agent orchestration (e.g. LangGraph) for complex workflows …)• Optimize LLM-based systems with DSPy
ML Theory & Research
• Analyze loss landscapes, generalization bounds, and scaling laws to guide model choices• Experiment with Mixture-of-Experts, sparse attention, and curriculum learning• Publish internal white-papers translating theory into concrete product roadmaps
Collaboration
• Translate energy-domain pain points into model specs with product & domain experts• Mentor engineers on LLM internals, prompt engineering, and cloud best practices
You may be a great fit if you have
· M.S. in CS, EE, Applied Math—or deep real-world expertise
· 5+ years building production ML systems, including:
o Distributed training of >7 B-parameter models with ZeRO-3 or Megatron-LM
o Serving quantized and pruned LLMs via Triton, vLLM, or TensorRT-LLM at scale
o Mastery of Python, plus willingness to dip into Go/Bash/TypeScript as needed
· Strong grasp of optimization theory, information theory, and alignment techniques
· A record of translating arXiv papers into hardened, customer-facing services (your GitHub repo comes handy here!)
· Clear communication of complex ideas to mixed technical audiences
· A builder’s bias for ownership, experimentation, and shipping
Even if you don’t tick every box, we’d still love to meet you. If you’re a 0-to-1 creator, relentless learner, and supportive teammate, you’ll thrive at Bright Fox AI.
We are an equal-opportunity employer. Diversity—in background, thought, and experience—makes us stronger and is essential to innovating for a brighter energy future.