ML Engineer
Content + Source + Freshness • 11 Dec 2025 • 95% confidence
Offer value
High value level due to advanced technical requirements and the potential to work on innovative AI projects with a talented team.
- Work on innovative AI projects with advanced technologies.
- Remote position fostering work-life balance.
- Collaboration with a dynamic and skilled team.
Pros
- Opportunity to work with cutting-edge AI solutions.
- Flexible remote working conditions.
- Chance to collaborate with skilled professionals in the field.
Cons
- High level of technical expertise required limiting applicant pool.
- Difficult tasks may be stressful for less experienced candidates.
- No salary information provided could be a deterrent.
Who it's for
Mid-level / Senior • Fully Remote (within Ukraine)
Good fit
- Mid and senior-level ML engineers.
- Candidates eager to innovate in AI technologies.
- Professionals with robust technical expertise in ML.
Not recommended for
- Entry-level professionals without AI experience.
- Candidates lacking a strong math or programming background.
- Those averse to complex project requirements.
Motivation fit
Key skills
About the job
We are looking for a passionate ML Engineer to implement AI solutions aimed at achieving business goals (CLAi automates a multitude of business operations simple and complex alike. That includes document reading, data entry, CRM management, calendar scheduling, and automated appointment booking.).
This role offers the opportunity to work on cutting-edge project and collaborate with a team of talented researchers and engineers in a stimulating and dynamic environment.
Minimum qualifications (must have hands-on experience with all of the below):
- LLMs (generation, alignment, extraction)
- Supervised fine-tuning (SFT) on ≥7B models with PEFT (LoRA/QLoRA) and full fine-tunes.
- Preference optimization (DPO/ORPO/PPO/RLAIF), rejection sampling, reward-model training.
- Constrained/structured decoding (regex/CFG/JSON-Schema), logit biasing, n-gram blocking, spec decoding.
- Distillation/compression (teacher-student, pruning, quant-aware training).
- RAG modeling side: dual-encoders, cross-encoders/rerankers (e.g., ColBERT/SPLADE), hard-negative mining.
Speech (ASR/diarization/VAD; optional TTS):
- Fine-tuning Conformer/Transducer/CTC or Whisper/wav2vec2/HuBERT on domain audio, incl. streaming/chunking.
- Robust segmentation (VAD), diarization (x-vectors/ECAPA-TDNN), punctuation & inverse text normalization.
- Data augmentation (SpecAugment, speed/tempo, noise/reverb), forced/CTC alignment and lexicon handling.
Data & evaluation:
- Large-scale corpus building: language ID, dedup/near-dedup (LSH/MinHash), toxicity/PII filters, perplexity/quality filters.
- Golden sets & adversarial suites; metrics for WER/CER, entity F1, extraction validity, factuality/hallucination, helpful-harmlessness.
- Reproducible experiments: seeds, checkpoints, ablations, learning-curve analysis, compute budgeting; crisp experiment reports.
Systems for training (efficiency)
- Distributed/memory-efficient training with FSDP/DeepSpeed ZeRO, gradient checkpointing, mixed precision, packing/bucketing, sequence-length curricula.
- Dataset pipelines: HF Datasets, WebDataset, streaming Parquet/TFRecords; tokenizer optimization and dataset QA.
Core foundations
- Solid math (linear algebra, probability, optimization) and ability to reason about loss design and bias/variance.
- Expert PyTorch (or JAX), custom modules/losses, profiling (cProfile/torch.profiler), multi-GPU runs.
- Habit of rigorous evals with automated harnesses and regression gates.
Nice to have
- Preference optimization at scale (PPO/DPO), safety classifiers.
- Quantization (GPTQ/AWQ/INT8) with minimal quality loss; ONNX/TensorRT; Triton/CUDA kernels.
- Multilingual modeling, phonetic/lexicon work for low-resource accents.
- Active learning & data programming (cleanlab/Snorkel).
