![]() ![]() Some technologies used in this project Architecture ![]() Docker - Deployment in Google Cloud Run.Scrapy: Web Crawling and Scraping Framework.Data Science REST API that could be consume by the backend as a client.Database storing of category-labeled news articles.Automated news web crawling-scraping executed several times per day.You can see the DataScience documentation here: Features You can see the project's Notion documentation here: My intention in this post is to show how I developed a complete automated news web scraper using Python and the Google Cloud Platform. I had the opportunity to work with great people in this particular project, that allowed us to build a very nice application in just 2 weeks. My team and I developed AlwaysUpdate which is a digital newspaper that automatically compiles daily news from different news portals, does a sentiment analysis of the text, and classifies its content as positive, neutral, or negative.Īs the project's Data Scientist, I developed the automated web crawler-scraper, and the sentiment analysis functionality, all served with a FastAPI REST API deployed in Google Cloud Platform. Photo: Suzy Hazelwood, the end of the intensive stage of the Platzi Master Program, the students have to develop a group project that involves several disciplines: Frontend-Backend development, and DataScience. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |