Index finger pointing up icon

From: Poland

Location: Katowice, PL

On Useme since 15 June 2020

Business bag icon

Deals 6

SUCCESSFUL: 6

DISPUTED: 0

FAILED: 0

About me

I am a Python developer with many years of experience. I offer scraper bots to scrape data from static and dynamic websites. In some cases also after logging. In addition, I deal with a text file operations, from cleaning files with duplicate / redundant data, to converting file extensions, other amongst I support these file extensions: .json, .csv, .xlsx, .txt, .log, and much more, for your special requests please contact.

Skills

Back office Bots Data analysis Databases Excel Python Sql

Portfolio

Portfolio item
Web scrapper olx.pl, gumtree.pl, otodom.pl

As the first stage of work, the application searches through advertisements that have appeared since the previous scrape. Ads are searched using the default location data serviced by websites. Then, after obtaining all the newly available offers,...

Portfolio item
Transformacja pliku txt do formatu excel

The assumption of the project was to extract data from a text file and standardize them to the Excel.

Portfolio item
Web scraper gratka.pl, sprzedajemy.pl, licytacje.k

Building a scraper bot to download data from websites: gratka.pl, sprzedajemy.pl, licytacje.komornik.pl. Everything connected to the PostgreSQL database and data optimization mechanisms.

Portfolio item
Integracja scrapera z rejestrem teryt

Integration of the scraper with the teryt register, which enables the normalization of the location to unified data.

Completed offers 1

  • Zlece wykonanie crawlera do zbierania ofert nieruchomosci. Zalozenia logiki: * crawler dodaje oferty do bazy danych * crawler nie dubluje ofert - rozpoznaje juz dodane * crawler normalizuje dane z oferty w zakresie * podzial administracyjny (wojewodztwo, powiat, gmina, miasto) * adres * typ nieruchomosci - dom/mieszkanie/dzialka (rolna, budowlana, lesna etc) * typ ogloszenia (sprzedaz, wynajem) * rodzaj ogloszenia (prywatne, posrednik) * link do strony * zdjecia * data utworzenia ogloszenia * Identyfikator ogloszenia * cena * powierzchnia * dane dodatkowe (liczba pokoi, pieter) * opis * crawler dotyczy 2 stron z ogloszeniami (strony do uzgodnienia podczas nawiazania kontaktu) * crawler wykrywa blokade stron lub blad serwera Zalozenia techniczne: * crawler posiada implementacje page object pattern * crawler jest uruchamiany w ramach pipelines (azure lub aws) * crawler jest uruchamiany 2x dziennie