AI Systems & Fine-Tuning Engineer (Python / Open-Source LLMs)

Please login or register as jobseeker to apply for this job.

TYPE OF WORK

Full Time

SALARY

500-2500

HOURS PER WEEK

TBD

DATE UPDATED

Mar 23, 2026

JOB OVERVIEW

About the role
We are looking for an AI Engineer to build, train, and optimize high-performance local models. This isn’t a "prompt engineering" or API-wrapping role—you’ll be working under the hood with fine-tuning, LoRAs, and deploying models directly on GPU hardware. Your goal is to take raw open-source weights and turn them into specialized, production-ready engines that are fast, efficient, and private.

What you’ll do
Specialized Training: Execute fine-tuning on local LLMs (LoRA, QLoRA, and full fine-tuning) to adapt models for specific domain tasks.

Model Optimization: Work with the latest open-source architectures (Qwen, Llama, Mistral) to maximize their utility in real-world applications.

Data & Pipelines: Build the "factory" for our models—handling everything from data scraping and cleaning to automated training and evaluation loops.

Inference Engineering: Optimize for speed. You’ll implement quantization, batching, and high-efficiency runtimes (vLLM/TGI) to keep latency low.

On-Prem Deployment: Manage model stability on local GPU servers and multi-GPU setups, ensuring high availability for internal services.

Rapid R&D: Stay at the bleeding edge—test new papers, quantization methods, and training techniques as soon as they hit GitHub.

What we’re looking for
Core Python: Deep experience building the scripts and tooling that power AI pipelines.

Fine-Tuning Expertise: Practical experience using PEFT, LoRA, and QLoRA to steer model behavior.

Local Execution: You’ve spent significant time running models on your own hardware or dedicated instances (not just hitting an OpenAI endpoint).

Architecture Knowledge: A strong handle on tokenization, context management, and preparing high-quality datasets.

The AI Stack: Fluency in Hugging Face, PyTorch, and the Transformers library.

Hardware Proficiency: Experience navigating CUDA, managing VRAM constraints, and orchestrating multi-GPU environments.

Problem Solver: You can debug a failing training run and understand why a model is "hallucinating" or underperforming.

Nice to have
Generative Media: Experience with ComfyUI, Stable Diffusion, or video generation workflows.

Edge Inference: Knowledge of GGUF, AWQ, or running models via engines like Ollama.

Scale: Experience with distributed training or managing massive, multi-terabyte datasets.

Agents: A background in building autonomous AI agents or internal productivity tools.

Top 3 Skills to select on OnlineJobs for this role:
Python (Absolute must-have for the codebase).

Machine Learning (To capture the general AI talent pool).

PyTorch (This filters for people who understand the underlying math/frameworks).

SKILL REQUIREMENT
VIEW OTHER JOB POSTS FROM:
SHARE THIS POST
facebook linkedin
  BENCHMARKS  
Loading Time: Base Classes  0.0010
Controller Execution Time ( Jobseekers / Job )  0.0133
Total Execution Time  0.0151
  GET DATA  
No GET data exists
  MEMORY USAGE  
1,499,856 bytes
  POST DATA  
No POST data exists
  URI STRING  
jobseekers/job/AI-Systems-Fine-Tuning-Engineer-Python-Open-Source-LLMs-1608201
  CLASS/METHOD  
jobseekers/job
  DATABASE:  onlinejobs (Jobseekers:$db)   QUERIES: 13 (0.0066 seconds)  (Hide)
0.0003   SELECT *
                                
FROM exrates
                                WHERE rate_name 
'USD-PHP' 
0.0004   SELECT *
FROM `employer_jobs`
WHERE `job_id` = 1608201
 LIMIT 1 
0.0004   SELECT *
FROM `employers`
WHERE `employer_id` = 917044
 LIMIT 1 
0.0007   SELECT COUNT(*) AS `numrows`
FROM `t_thread` `t`
LEFT JOIN `t_thread_misc` `miscON `t`.`id` = `misc`.`thread_id`
WHERE `t`.`job_id` = 1608201
AND `misc`.`idIS NULL 
0.0004   SELECT e.business_namee.logoe.websitee.rebill_datee.date_added member_datehitsDATEDIFF('2026-04-21',ej.date_added) duration_daysDATEDIFF('2026-04-21',e.rebill_date) duration_rebillej.*, e.deactivate FROM employers eemployer_jobs ej WHERE e.employer_id ej.employer_id AND
                                   ((
e.user_level >= '500' AND ej.date_added <= e.rebill_date)
                                   OR 
e.employer_id '' OR (ej.date_approved <> '2000-01-01' and DATEDIFF('2026-04-21',ej.date_added) <= 14 ))
                                   AND 
e.deactivate != AND ej.deleted AND job_id '1608201' 
0.0003   SELECT *
FROM `employer_jobs_skills` `ejs`
LEFT JOIN `skills_categories` `scON `ejs`.`skill_id` = `sc`.`id`
WHERE `job_id` = 1608201 
0.0015   UPDATE employer_jobs SET hit_counts '***Mar-23-2026=170***Mar-24-2026=35***Mar-25-2026=35***Mar-26-2026=12***Mar-27-2026=8***Mar-28-2026=4***Mar-29-2026=4***Mar-30-2026=11***Mar-31-2026=27***Apr-01-2026=2***Apr-02-2026=4***Apr-04-2026=1***Apr-05-2026=2***Apr-06-2026=4***Apr-07-2026=5***Apr-08-2026=1***Apr-10-2026=1***Apr-11-2026=1***Apr-12-2026=2***Apr-13-2026=3***Apr-14-2026=1***Apr-15-2026=1***Apr-16-2026=2***Apr-17-2026=1***Apr-21-2026=1' WHERE job_id'1608201'  
0.0006   UPDATE employer_jobs SET monthly_hits '***Mar-2026=306***Apr-2026=32' WHERE job_id'1608201'  
0.0009   SELECT date_sent FROM jobseeker_sent_emails WHERE jobseeker_id '' AND job_id '1608201' AND status LIKE 'sent%' ORDER BY id DESC  
0.0003   SELECT *
FROM `employer_jobs_skills` `ejs`
LEFT JOIN `skills_categories` `scON `ejs`.`skill_id` = `sc`.`id`
WHERE `job_id` = 1608201 
0.0003   SELECT COUNT(*) AS `numrows`
FROM `employer_jobs`
WHERE `employer_id` = '917044'
AND `date_added` >= '2022-06-08' 
0.0003   select from teasers 
0.0002   SELECT FROM skill_categories WHERE skill_cat_id='' 
  HTTP HEADERS  (Show)
  SESSION DATA  (Show)
  CONFIG VARIABLES  (Show)