Crawl data from world bank according to indicator names.

Suljettu Julkaistu 2 kuukautta sitten Maksettu toimituksen yhteydessä
Suljettu Maksettu toimituksen yhteydessä

I am looking for a freelancer to write python program, to crawl data from the World Bank based on specific indicator names. The indicator names will be provided as arguments of the input python.

1. Should have a arguments of Proxy IPs and use random IP address according to the given IPs.

2. Should retry 3 times to craw data

Passed arguments:

(1)--config [login to view URL], including indicator name lists, may be json format, to specify global year range and specify indicator year range. For example:

{

"global": {

"starttime": "2021-01-01", //YYYY-MM-DD, if not specify, no limit to starttime

"endtime": "2021-01-01", //YYYY-MM-DD, if not specify, no limit to endtime

"craw_interval": 3000, //ms, craw interval between each

"proxies": [ //support multiple proxies, use random proxy. If proxy connect failed, try 3 times.

{

"proxy_type": "http",

"proxy": "[login to view URL]"

}

]

},

"indicators": [

{

"meta_name": "[login to view URL]",

"indicator_name": "Population Ages",

"indicator_code": "pop",

"starttime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

"endtime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

},

{

"meta_name": "[login to view URL]", //origin indicator name of world bank

"indicator_name": "Population Ages",

"indicator_code": "pop",

"starttime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

"endtime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

}

],

"mapping_dict": {

"country": [

{

"name": "USA",

"code": "USA",

"alias": [ // search every alias ignore case.

"United States of America",

"US"

]

}

],

"field": [ //use to remapping the output values when a single indicator contains multiple values fields. default field is value

{

"meta_name": "value", // meta_name is the original field name. the key of "value" represent original default value field.

"meta_value": "value"

},

{

"meta_name": "score", // meta_name is the original field name. the key of "value" represent original default value field.

"meta_value": "Final Score"

}

]

}

}

(2)--output output path of the final json. The output format should be json. Output format:

{

"indicators": [

{

"meta_name": "[login to view URL]",

"indicator_name": "Population Ages",

"indicator_code": "pop",

"starttime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

"endtime": "2021-01-01", //YYYY-MM-DD, if not specify, follow global config

"status": "success", // if crawl failed, set status = failed.

"errmsg": "OK", //if crawl failed, show error message here

"total": 2939293, //total data entry counts

"countries": 238823, //total country counts

"years": 60 //total years

},

],

"data": [

{

"datasource": "worldbank",

"ref_link": "[login to view URL]", //the indicator's original link

"meta_name": "[login to view URL]",

"indicator_name": "Population Ages",

"indicator_code": "pop",

"country_name": "USA",

"country_code": "USA",

"crawl_time": "2022-01-01 12:00:00",

"year": "2022",

"starttime": "2022-01-01", // if data source only contains year like 2022, set it to start of the year

"endtime": "2022-12-30", // if data source only contains year like 2022, set it to end of the year

"values": { // use field mapping to convert to new field name first. the key should be the new field name.

"value": 123,

"Final Score": 300

}

}

]

}

3. Output should contains

(1) failed indicator names and failed reason

(2) remapping indicator_name and value fields

[login to view URL] provide

(1)[login to view URL]

(2)Python code

(3)A simple test case([login to view URL] and test command)

Skills and Experience:

- Strong experience in web scraping and data crawling

- Proficiency in Python or another suitable programming language for web scraping

- Familiarity with the World Bank's data structure and API

Data Format:

- The crawled data should be in JSON format.

Data Cleaning and Structuring:

- The client requires the data to be cleaned and structured according to specific data attributes.

Please provide examples of similar projects you have completed in the past.

Python Web Crawling

Projektin tunnus: #37219018

Tietoa projektista

23 ehdotusta Etäprojekti Aktiivinen kuukausi sitten

23 freelanceria on tarjonnut keskimäärin $189 tähän työhön

chirgeo

Hi! I'm George, a web crawler expert with many successful data crawling projects in my portfolio. I'm confident I can provide you with the necessary expertise to meet your requirements for this job. I have the necess Lisää

$500 USD 3 päivässä
(127 arvostelua)
7.7
mananraja

hi I can provide you Python script as well as data in JSON data, according to the format you shared, from World Bank website. I can start right away. Abdul H.

$150 USD 1 päivässä
(147 arvostelua)
6.8
schoudhary1553

Top 1% in Freelancer.com Hi, Greetings! ✅checked your project details: ✅Completed Time: In project deadline We have worked on 900 + Projects. I have 6 + years of the experience in same kind of projects. If you are look Lisää

$180 USD 3 päivässä
(92 arvostelua)
6.9
sonus21

Hello Suley! I hope you're well. I'm a senior Scraper developer with experience in developing scraper using scrappy and headless browsers. I can deal with bypassing IP throttling limit, ban and captcha solve, storing t Lisää

$200 USD 7 päivässä
(54 arvostelua)
6.2
manojguragain184

World bank scraper in python. Dear, Client I am thrilled to express my interest in your web scraping and automation project. I am confident in my ability to deliver outstanding results that align with your requirement Lisää

$250 USD 3 päivässä
(9 arvostelua)
5.0
gm341473

Dear sir, I have gone your project, and it matches my expertise. I'm confident of performing your project work. I have 6+ years of experience in Web scraping, Python, Debugging, Software architecture. I can do your pro Lisää

$140 USD 2 päivässä
(18 arvostelua)
5.1
Techalchemy

Hi Sir, As a highly skilled and experienced, I am confident that I can provide the high-quality work you need. I am ready to start the work right away. Thank you.

$140 USD 7 päivässä
(8 arvostelua)
4.4
Demenntor

Dear Client, I am excited to submit my proposal for your project, which involves writing a Python program to crawl data from the World Bank based on specific indicator names. I have extensive experience in web scrapin Lisää

$220 USD 7 päivässä
(16 arvostelua)
3.7
Dhawan8787

Hi sir, I'm excited about your project and confident in my ability to deliver your project . I'm committed to exceeding your expectations and ready to start from right away . Thank you .

$140 USD 7 päivässä
(4 arvostelua)
3.4
tkmaina868

I am professional Swift coder with skills including Web Crawling and Python. Please contact me to discuss more about this project. Thank you

$100 USD 3 päivässä
(1 arvostelu)
3.4
DataScinceFizer

Hello, I am DataScinceFizer I would like to work on using Python to create Automated Scripts Scraping tools. I have extensive experience in Python to create website scraping tools to collect data from your target websi Lisää

$80 USD 2 päivässä
(4 arvostelua)
2.1
narukaconsultan5

Hello, I am writing to bid on your project to write a Python program to crawl data from the World Bank based on specific indicator names. I have strong experience in web scraping and data crawling, and I am proficient Lisää

$220 USD 7 päivässä
(4 arvostelua)
1.8
contentnikunj

Hello there! I understand you are looking for a freelancer to write a python program that will crawl data from the World Bank based on specific indicator names. I believe I am the best fit for this project because of m Lisää

$99 USD 1 päivässä
(2 arvostelua)
0.0
Aliyanpops

Hello, I have the web scraping skills that you require to attain the data you need from the World Bank. To further create an understanding, I would suggest that you share a portion of the project detail so that I can f Lisää

$100 USD 5 päivässä
(0 arvostelua)
0.0