As a key member of the AI Services Division, the AI Infrastructure Engineer will be responsible for designing, deploying, and maintaining high-performance AI computing environments.
The role focuses on NVIDIA GPU-based systems and requires deep technical expertise in AI infrastructure and server technologies.
The ideal candidate will combine strong technical acumen with hands-on experience in system implementation and customer engagement.
Responsibilities
1.
Pre-Sales Technical Support
Provide pre-sales technical consultation for AI servers and related hardware solutions, including requirements analysis, solution design, performance evaluation, and technical communication.
Understand customers' AI application scenarios (e.g., large model training, inference, HPC computing, and data analytics) to recommend optimal server configurations and system architectures.
Collaborate with the sales team to prepare technical proposals, bidding documentation, solution presentations, and demonstrations.
Participate in Proof of Concept (POC) testing, performance tuning, and solution validation to ensure alignment with customer business objectives.
2.
Deployment Support
Physical Deployment:
Perform onsite installation of AI infrastructure, including server setup, network integration, cabling, and labeling.
Logical Configuration:
Configure systems and networks, conduct validation tests, and benchmark performance.
Operational Enablement:
Deliver system configuration and operation documentation.
Implement foundational operational tools (e.g., monitoring and alerting systems).
3.
Maintenance and Support
Diagnose and troubleshoot hardware and software issues in AI infrastructure.
Perform hardware replacement, software updates, or implement workarounds as required.
Escalate complex technical issues to vendors for advanced support when necessary.
Qualifications and Skills
Bachelor's degree or higher in Computer Science, Electronics, Communications, Information Engineering , or a related discipline.
5+ years of experience in server OEM environments or related fields such as AI computing, big data, or HPC infrastructure .
Strong understanding of server hardware architectures (CPU/GPU, memory, storage, networking).
Hands-on experience deploying or maintaining large-scale NVIDIA AI infrastructure , including at least one of the following: B200, GB200, or GB300 .
Familiarity with major computing platforms such as NVIDIA, Intel, and AMD .
Excellent problem-solving, collaboration, and customer service skills.
Proficient in English (reading and writing); able to interpret technical documentation and communicate effectively.
Preferred Attributes
Experience working with AI infrastructure deployment at scale in enterprise or data center environments.
Knowledge of Linux system administration, performance tuning , and monitoring tools .
Strong analytical mindset and eagerness to learn emerging AI technologies.Would you like me to optimize this further for a specific audience (e.g., for a LinkedIn job post, internal HR use, or a proposal to a client)?
The tone and structure can be fine-tuned accordingly.