Company Description
Freshworks creates an effortless and rapid way for corporations to impress their customers and workers. By applying a novel method to the creation and supply of software, which is reasonably priced, swiftly implemented, and designed with the end user in mind. Based in San Mateo, California, Freshworks operates a global team from 13 global locations serving over 65,000 companies - from startups to listed firms – whose dependability on Freshworks' software-as-a-service boosts their customer (CRM, CX) and employee experience (ITSM).
Freshworks’ assortment of cloud-based software includes Freshdesk (multi-channel customer support), Freshsales (sales automation), Freshmarketer (marketing automation), Freshservice (IT help desk), and Freshchat (AI-powered bots), all reinforced by Neo, our base platform of shared services.
Freshworks has been featured in international national press including CNBC, Forbes, Fortune, Bloomberg and has been recognized as a great place to work in San Francisco and Denver by BuiltIn for the last three years. Our customer ratings have awarded Freshworks products the highest TrustRadius Software ratings and numerous G2 Best of Awards for Best Feature Set, Best Value for the Price, and Best Relationship.
Job Description
The fundamental responsibilities of the position include:
- Engineer and construct a real-time data pipeline for Data intake for real-time business use cases.
- Design sophisticated and efficient functions to transform raw data sources into robust, trustworthy elements of our data lake.
- Expand our analytics abilities with quicker, more dependable data pipelines, and advanced tools, managing petabytes of data daily.
- Innovate and develop new platform features, aiding our bid to make data accessible to cluster users in every format, with low latency and scalable output.
- Implement changes to our data platform, restructuring/designing as required and diagnosing any issues across the technical stack completely.
- Think creatively to execute solutions with new components and various new technologies on AWS, and Open Source for the successful completion of various projects.
- Improve and optimize existing features or data processes for performance and stability.
- Write unit tests and facilitate continuous integration.
- Have a keen eye for quality ensuring minimal production downtimes.
- Mentor colleagues, disseminate information and knowledge, and contribute to a strong team.
- Monitor job performances, manage file system/disk-space, cluster and database connectivity, keep track of log files, manage backup/security, and provide solutions to various user issues.
- Cooperate with multi-functional and business teams.
Qualifications
We seek a candidate with
- demonstrated experience in a Big Data Engineering role with on-the-job expertise in Apache SparkTM (either Scala or PySpark Preferred) and associated performance optimization.
- Advanced practical knowledge of SQL and a comfortable familiarity with a range of databases.
- Practical knowledge of various API interfaces for Bulk or Stream-based data extraction and load processes is imperative.
- Experience creating and implementing a range of data engineering pipelines into production, including using automation best practices for CI/CD.
- Experience conducting root cause analysis on all data and processes to answer specific questions and identify areas for improvement.
- Knowledge on processes supporting data transformation, data structures, metadata, dependency and workload management.
- A successful track record of manipulating, processing, and extracting value from large disconnected data sets.
- Practical knowledge of Kafka, Spark, stream processing, and scalable 'big data' data stores.
- Experience with cloud solutions on top of AWS
- Beneficial to have ML-ops Knowledge
- Preferred Experience: 3-5 Years
Additional Information