Location: USA / Job Type: Full Time
Please email jobs@purgo.ai
The Data Architect role offers the successful candidate the opportunity to pioneer the adoption of generative AI in design, development and migration of data applications. This is a hands-on technical role involving deep collaboration with both Purgo AI’s product/engineering team and its customers/partners. The role drives the maturation and adoption of Purgo AI’s redefined software design lifecycle amongst both cloud data warehouse partners and customers.
Purgo AI is an application design studio powered by generative AI, which interprets high-level business problem statements and generates design, source code and deployment of data applications over cloud data warehouses. The product’s fine-tuned LLMs specialize in solving business problems with business intelligence (ETL/ELT), cloud migration and on-demand machine-learning & inferencing for forecasting, anomaly detection & pattern recognition. Purgo AI integrates out-of-box with all leading cloud data warehouses including Databricks, Snowflake, Microsoft Fabric, Google BigQuery and AWS RedShift.
• Business Intelligence (ETL/ELT): Business analysts feed new ETL/ELT user requirements through Purgo AI’s Jira app, which trigger the generation of behavior-driven design (BDD) requirements for solving the Jira tickets and test harnesses for providing QA. Purgo AI generates source code from integrated code-generation LLMs by using the BDD specifications without needing any human prompting. The generated code is subject to pre-generated quality assurance tests and test fails re-trigger generation of the source code. The final source code is ready for end deployment over the cloud data warehouse after any inspection/approval by the business analyst team. The entire process has end-to-end traceability through Purgo AI generated log entries across Jira and GitHub systems.
• Cloud Migration: Transformation teams interpret legacy data applications through Purgo AI’s GitHub plug-in, which triggers the generation of a cloud migration plan for stored procedures, data schema and data records for a specific target cloud data warehouse. The migration plan includes behavior-driven designs for each of the components, which are used to generate source code from integrated code-generation LLMs without needing any human prompting. Purgo AI leverages the generated migration plan for executing a methodical migration all the way to data migration. Purgo AI tests the migration with pre-generated tests that validate data, table relationships, procedures and data schema. The product is addressing a trillion-dollar legacy application market with this end-to-end solution.
• On-demand Machine-learning and Inferencing: Purgo AI interprets predictive business problem statements and generates the design for training an on-demand machine-learning (ML) model along with necessary training/test datasets. Purgo AI integrates with ML scripting frameworks like MLFlow and TensorFlow to execute training, testing and deploying new ML models custom-designed from the problem statement in the form of notebooks deployed over the cloud data warehouse. Purgo AI infers from the deployed model(s) by using the parameters described in the problem statement to generate a predicted solution for the business user. The product is capable of delivering this on-demand machine-learning and inferencing for business problems across forecasting, anomaly detection and pattern recognition use-cases.
The product integrates seamlessly with Jira, Github (to interpret existing/legacy source code), code LLMs (from Github, AWS, Mistral and Meta), and test automation platforms (like Selenium, Pytest etc.). The company’s co-founder and CTO, Sang Kim, has been an engineering leader across VMWare and Blackberry. The company is based in Palo Alto, CA and co-created by The Hive, a venture studio focused on data and AI in the enterprise.
Please email jobs@purgo.ai.
Purgo AI is an affirmative action employer and welcomes candidates who will contribute to the diversity of the company.