Why Work at Lenovo
We are Lenovo. We do what we say. We own what we do. We WOW our customers.
Lenovo is a US$57 billion revenue global technology powerhouse, ranked #248 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world's largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo's continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY).
Want more jobs like this?
Get jobs in Hong Kong delivered to your inbox every week.
This transformation together with Lenovo's world-changing innovation is building a more inclusive, trustworthy, and smarter future for everyone, everywhere. To find out more visit www.lenovo.com, and read about the latest news via our StoryHub.
Description and Requirements
Summary
Data Asset team is looking for a data engineer with a solid technical background to build and support cloud based data asset platform. Our platform enables internal engineers and analysts to process SQL or transformations using Apache Spark, Python on compute for batch, streaming and ML processing, executing over schemaed data stored in data lake and data warehouse. Our team manages data ingestions from various data sources, build and support ETL pipelines, and continuously look for solutions to enhance the robustness and scalability of the data architecture.
The Job
You will be responsible for application development and ongoing maintenance support of the Data Analytics area with the following scope of duties:
- Manage cloud resources including EC2, EMR, S3, Redshift, Glue, etc.
- Design data schema and operate internal data warehouses and SQL/NOSQL database systems
- Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using SQL and cloud big data technologies
- Build robust and scalable data integration (ETL) pipelines using SQL, PySpark and cloud data services like EMR, Redshift, etc.
- Implement suitable data model for data warehouse in order to achieve efficient data retrieval
- Gather business and functional requirements and translate these requirements into robust, scalable, operable solutions with a flexible and adaptable data architecture.
- Collaborate with other Engineers and Product Managers to recognize and help adopt best practices in data engineering and analytics: data integrity, test design, analysis, validation, and documentation
The Person
- 5+ years working experience in software development, data engineering, business intelligence or related fields
- 2+ years data modeling, ETL development, and data warehousing experience
- 2+ years of experience using big data technologies (Hadoop, Spark, Hive, Impala, etc.)
- Good understanding of cloud infrastructure
- Advanced SQL and query performance tuning skills
- Experience with one of the functional scripting languages (Python, Scala, Java etc.) to process semi-structured or unstructured data inputs.
- Knowledge of professional software engineering practices & best practices for the full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations.
- Strong customer focus, ownership, urgency, and drive.
- University graduate in Computer Science, Information Technology, Data Science or related disciplines
- Effective verbal and written communication in English and Chinese
- Good team player with strong interpersonal and communication skills
- Forward thinking, embracing changes, fast-learning, energetic, able to perform in challenging environment
- Peferred Qualifications
- Working experience with AWS or GCP big data technologies (S3, Redshift, EMR, Kinesis, DynomoDB, BigQuery, Data Flow)
- Experience with HBase, Kafka, Flink.
- Database design and administration experience with Oracle, Teradata or other large commercial relational database systems
- Working experience with NOSQL databases
- Experience contributing to the system design or architecture (architecture, design patterns, reliability and scaling) of new and current systems
- Proven success in communicating with users, other technical teams, and senior management to collect requirements, describe data modeling decisions and data engineering strategy.
- Experience providing technical leadership and educating other engineers for best practices on data engineering.
- Candidates with less experience will also be considered
#LPS
Additional Locations:
* Hong Kong