صاحب العمل نشط
حالة تأهب وظيفة
سيتم تحديثك بأحدث تنبيهات الوظائف عبر البريد الإلكترونيحالة تأهب وظيفة
سيتم تحديثك بأحدث تنبيهات الوظائف عبر البريد الإلكترونيJob Description
Mandatory Skills
Python Pyspark Apache Spark SQL Json Cloud
Additional Skills
Docker Kubernetes Communication skills
Key Responsibilities:
Designing and developing robust PySpark applications for largescale data processing.
Building and optimizing data ingestion transformation and storage processes.
Implementing efficient algorithms and data structures for distributed computing.
Collaborating with crossfunctional teams to integrate datadriven solutions into business processes.
Troubleshooting performance bottlenecks and ensuring high availability and reliability of data pipelines.
Writing and optimizing SQL queries for data extraction and manipulation.
Required Skills and Qualifications:
Bachelors/Masters degree in Computer Science Engineering or a related field.
Proven experience (310 years) in Python development with a focus on PySpark.
Strong understanding of distributed computing principles and experience with Apache Spark.
Proficiency in SQL and experience with relational databases (MySQL PostgreSQL etc.).
Experience with data serialization formats such as JSON Parquet Avro.
Familiarity with cloud platforms (AWS Azure GCP) and containerization technologies (Docker Kubernetes) is a plus.
Excellent problemsolving skills and ability to work independently or as part of a team.
Good communication skills with the ability to effectively collaborate with stakeholders.
Preferred Qualifications:
Certification in Apache Spark or related technologies.
Experience with stream processing frameworks like Apache Kafka.
Knowledge of machine learning frameworks (e.g. TensorFlow PyTorch) for data analysis.
Additional Information
Required Qualification
Master of Computer Applications (M.C.A.) Bachelor of Engineering Bachelor of Technology (B.E./B.Tech.)
دوام كامل