This repository contains detailed documentation and diagrams that describe the different components of the system and how they interact.
Prices Crawler system is built on top of MongoDB and comprises several applications that work together to provide a seamless user experience. The content-base application provides an abstract implementation, which is used by the content-api-example. This, in turn, is consumed by the web-app, mobile-app, and feed by catalog-data-consumer (implementation not provided).
In addition to these components, it also includes the process-worker application, which provides several workflows for ensuring data consistency and performing routine maintenance tasks.
This documentation is intended to provide a comprehensive overview of the system architecture and help developers understand the different components and their interactions. We hope that you find it useful and welcome any feedback or suggestions for improvement.
This System Architecture is a high-level overview of prices-crawler system that outlines the different components and how they interact with each other. It provides a comprehensive view of the system that is used to guide the development process and ensure that all components are properly integrated.
graph TB;
mongoDb(MongoDb);
contentBase(content-base - abstract);
contentApi(content-api-example);
webApp(web-app);
mobileApp(mobile-app);
processWorker(process-worker);
catalogConsumer(catalog-data-consumer);
mongoDb --> contentBase;
contentBase --> contentApi;
contentApi --> webApp;
contentApi --> mobileApp;
catalogConsumer --> contentApi;
contentBase --> processWorker;
processWorker --> ro2(clear-product-incidents-process<br>clear-product-incidents-automated-process<br>database-restore-process<br>product-data-consistency-process<br>product-data-update-process);