Opens intel.wd1.myworkdayjobs.com in a new tab
About This Role
- Intel's Data Center Network Edge AI team builds the AI software stack that powers next-generation workloads across Intel® Xeon® processors, Intel® dGPU based on Intel Edge AI platforms.
- We deliver the algorithms, frameworks, and optimizations that bring large-scale multi-modal intelligence to Intel customers worldwide.
- We are looking for an intern to contribute to our research and productization of Vision-Language Models (VLM) and Vision-Language-Action (VLA) models.
- You will work alongside senior software engineers to advance state-of-the-art multi-modal and embodied AI, while ensuring these models run efficiently on Intel hardware.
- Responsibilities Conduct applied research on VLM / VLA architectures, pre-training, fine-tuning, and alignment techniques (SFT, RLHF, DPO, GRPO); reproduce and extend recent work such as OpenVLA, RT-2, π05, PaLI-Gemma, Qwen-VL, and InternVL.
- Design and implement multi-modal data pipelines for cleaning, synthesis, and augmentation of image-text-action datasets.
- Investigate efficient fusion strategies between vision encoders (ViT, SigLIP, DINOv2) and language backbones (LLaMA, Qwen, Mistral), including connector design and visual token compression.
- Explore VLA-specific components such as action heads (discrete tokenization, diffusion policy, flow matching), long-horizon planning, and closed-loop control.
- Apply model optimization techniques — quantization (INT8 / FP8 / INT4, AWQ, GPTQ, SmoothQuant), pruning, distillation, KV-cache optimization, and speculative decoding — to enable efficient deployment on Intel platforms.
- Collaborate with infrastructure engineers to deploy and benchmark optimized models across Intel® Xeon®, Intel® Gaudi®, Intel® Arc™ GPU, and Intel® Core™ Ultra (NPU) targets.
- Evaluate models against industry benchmarks (MMBench, MME, SEED-Bench, LIBERO, SimplerEnv) and contribute to internal evaluation suites.
- Publish results in technical reports and, where appropriate, top-tier venues (NeurIPS, ICML, ICLR, CVPR, CoRL, RSS).
Requirements
- The candidate must have the right to work in the country of employment without restriction.
- Currently pursuing an MS or PhD in Computer Science, Electrical Engineering, Artificial Intelligence, Mathematics, or a related technical field.
- Available for a minimum of 3 months of full-time or near full-time engagement.
- Strong foundation in deep learning, including Transformers, diffusion models, and reinforcement learning.
- Proficiency in Python and PyTorch; familiarity with distributed training (DeepSpeed, FSDP, Megatron-LM).
- Demonstrated project or publication experience in at least one of the following: multi-modal large models, embodied AI/Lerobot learning, model compression and inference acceleration, or vision-language pre-training.
- Preferred Qualifications Hands-on experience with model quantization, pruning, or distillation, and inference frameworks such as Pytorch, vLLM, SGLang, TensorRT-LLM, or llama.cpp.
- Prior deployment experience on Intel platforms (Xeon, Arc, Core Ultra).
- Experience with robotics simulation environments (Isaac Sim, MuJoCo, ManiSkill, RoboCasa) or real-robot systems.
- Open-source contributions to relevant AI projects on GitHub.
Sourced directly from Intel’s career page
Your application goes straight to Intel.
Opens intel.wd1.myworkdayjobs.com in a new tab
Specialisation
Open roles at Intel
765 positions
Job ID
/job/PRC-Shanghai/AI-Software-Engineer-Intern_JR0283188-1
Get matched to roles like this
Upload your resume once. We’ll notify you when matching roles open up.
Join talent pool — freeSimilar Other roles
Samsung Semiconductor
Thermal Engineer
San Jose, California, United States|Other
Samsung Semiconductor
Senior Manager, OLED Field Applications Engineering
San Jose, California, United States|Other
Samsung Semiconductor
Compensation Partner
San Jose, California, United States|Other
Micron Technology
HVM PEE Bench Operation Equipment Technician (内製修理テクニシャン)
Hiroshima - Fab 15, Japan|Other