Post Job Free
Sign in

Generative AI Engineer

Company:
Bright Fox AI
Location:
India, PA
Posted:
May 20, 2025
Apply

Description:

Who We Are

Bright Fox AI is a team of cloud architects, machine-learning theorists, and energy veterans turning the latest research on large language models (LLMs) into production-grade digital employees.

About the Role

You’ll be an engineer operating where LLM research meets cloud-native scale. Your charter is to design and own the infrastructure that trains, fine-tunes, and serves massive models—then squeeze every watt of insight out of them for our energy clients.

If you can diagram a Transformer’s attention flow, debate LoRA vs. QLoRA compression, and spin up a GPU fleet with in cloud environments, come build the future with us.

In this role you will tackle in these areas:

Cloud & DevOps

• Architect multi-cloud (AWS / Azure / GCP) GPU platforms • Embed GitOps pipelines (ArgoCD, GitHub Actions) for end-to-end model lifecycle

LLM Systems

• Fine-tune and distill GPT-class models using DeepSpeed, FSDP, and PEFT (LoRA/QLoRA)• Build Retrieval-Augmented Generation (RAG) pipelines with vector DBs (FAISS, Weaviate, …)• Construct LLM applications with frameworks such as LangChain, LlamaIndex and multi-agent orchestration (e.g. LangGraph) for complex workflows …)• Optimize LLM-based systems with DSPy

ML Theory & Research

• Analyze loss landscapes, generalization bounds, and scaling laws to guide model choices• Experiment with Mixture-of-Experts, sparse attention, and curriculum learning• Publish internal white-papers translating theory into concrete product roadmaps

Collaboration

• Translate energy-domain pain points into model specs with product & domain experts• Mentor engineers on LLM internals, prompt engineering, and cloud best practices

You may be a great fit if you have

· M.S. in CS, EE, Applied Math—or deep real-world expertise

· 5+ years building production ML systems, including:

o Distributed training of >7 B-parameter models with ZeRO-3 or Megatron-LM

o Serving quantized and pruned LLMs via Triton, vLLM, or TensorRT-LLM at scale

o Mastery of Python, plus willingness to dip into Go/Bash/TypeScript as needed

· Strong grasp of optimization theory, information theory, and alignment techniques

· A record of translating arXiv papers into hardened, customer-facing services (your GitHub repo comes handy here!)

· Clear communication of complex ideas to mixed technical audiences

· A builder’s bias for ownership, experimentation, and shipping

Even if you don’t tick every box, we’d still love to meet you. If you’re a 0-to-1 creator, relentless learner, and supportive teammate, you’ll thrive at Bright Fox AI.

We are an equal-opportunity employer. Diversity—in background, thought, and experience—makes us stronger and is essential to innovating for a brighter energy future.

Apply