What you will be working on: - Developing a highly-scalable distributed computing framework - Implementing various distributed algorithms such as model/data parallel frameworks, parameter servers, dataflow based asynchronous data communication - Developing a MPI-like communication framework for Deep Learning - Developing distributed memory constructs - Ensuring optimal performance: balance job distribution, minimize communication, balance computation vs. communication - Optimizing code for various computing hardware backends - Interacting with deep learning researchers to support different distribution frameworks - Interacting with driver developers and hardware architects to influence low level features
Inside this Business Group
Who we are looking for: - Master's or Ph.D. in computer science, electrical engineering or related fields (statistics, applied math, computational neuroscience) - Strong skills in algorithms - Strong programming skills (C++ and Python preferred) - Experience writing software for GPUs, HPC or distributed systems is preferred - Skills or experience in quantitative areas such as machine learning and computer vision is preferred
Intel Nervana, leveraging Intel’s world leading position in silicon innovation and proven history in creating the compute standards that power our world, is transforming Artificial Intelligence (AI). Harnessing silicon designed specifically for AI, end-to-end solutions that broadly span from the data center to the edge, and tools that enable customers to quickly deploy and scale up, Intel Nervana is inside AI and leading the next evolution of compute.