$$$
{{ $t($store.state.user.experience_value_in_dollars) }}
Senior
{{ $t($store.state.user.experience_search_name) }}
0
jobs
Python Developer - Web/Data Scraping
ABHIRAM KANNURI
,
London, United Kingdom
Experience
Skills
I'm offering
Highly accurate and experienced Data Analyst adept at collecting, analyzing, and interpreting large datasets, developing new forecasting models, and performing data management tasks. Processing extensive analytical skills, strong attention to detail, and a significant ability to work in team environments, Abhiram is presently looking for a Machine Learning Engineer, Business Analyst, Data Scientist, Data Analyst, and Data Engineers position with a forward-moving company.
I do:
Web Scraping using Python Libraries(Scrapy, BeautifulSoup)
Building ETL Pipelines in Python
Create data workflows on KNIME
Processing for unstructured and unconventional data in Python
I do:
Web Scraping using Python Libraries(Scrapy, BeautifulSoup)
Building ETL Pipelines in Python
Create data workflows on KNIME
Processing for unstructured and unconventional data in Python
Markets
United States
(Remote
only)
Canada
(Remote
only)
United Kingdom
(Remote
only)
Ireland
(Remote
only)
Lithuania
(Remote
only)
Rest of Euro area
(Remote
only)
Denmark
(Remote
only)
Norway
(Remote
only)
Sweden
(Remote
only)
Finland
(Remote
only)
Links for more
Once you have created a company account and a job, you can access the profiles links.
Industries
Language
English
Fluently
Available
My experience
2021 - ?
job
Python Data Engineer
Talysys Ltd UK.
york, UK
● Web Scraping using Python Libraries(Scrapy, BeautifulSoup)
● Building an ETL Pipelines in Python
● Create data workflows on KNIME
● Processing for unstructured and unconventional data in Python.
● Web Scraping using Python Libraries(Scrapy, BeautifulSoup)
● Building an ETL Pipelines in Python
● Create data workflows on KNIME
● Processing for unstructured and unconventional data in Python.
Python, ETL, Web, Processing, Knime
2021 - 2021
project
Data Scraping from IKEA Site
Scraping all products list from www.ieka.in
Middlesex University
Python, Sql, Webscraping, API
2021 - 2021
project
As a Python expert for modification to YAML parser
custom built Python module which is used to parse and validate a YAML file. The module has methods to retrieve properties from the inputted YAML file but the input file is now too bloated and contains a lot of duplicated fields across the different array elements so I would like to add a "global" section to house those common properties
Python, Sql, API, Data analysis and data capturing
2021 - 2021
project
Python script to load json data files into postgresDB
Loading Data into DB
PostgreSQL, Sql, Python
2021 - 2021
project
Scraping social profiles for emails
A spreadsheet containing a list of Instagram handles (other populated columns are irrelevant to this task). For each handle, the corresponding Instagram profile -- with very high likelihood -- contains an email address in the text field description. Visit the page corresponding to each profile and extract
-email address (if available)
-Instagram account name
-first name
-last name (if existing)
and populate the corresponding columns.
We will need some logic to parse and distinguish first names from last names. I will revise this shortly, but if it helps:
if only one word, call it the first name.
if greater than or equal to 1 word,
if the first-word length is less than or equal to 3, call words 1 and 2 the first name and any extra word(s) the last name.
-email address (if available)
-Instagram account name
-first name
-last name (if existing)
and populate the corresponding columns.
We will need some logic to parse and distinguish first names from last names. I will revise this shortly, but if it helps:
if only one word, call it the first name.
if greater than or equal to 1 word,
if the first-word length is less than or equal to 3, call words 1 and 2 the first name and any extra word(s) the last name.
Python, API, Social Media
2020 - 2021
internship
Junior Data Scientist Intern
SanSys Technologies.
UK
● Conduct software analysis, programming, testing and debugging.
● Building algorithms and models to address business problems
● Presenting information using data visualization techniques
● Design, implement and maintain the software using Data Science
● Take part in software and architectural development activities.
● Conduct software analysis, programming, testing and debugging.
● Building algorithms and models to address business problems
● Presenting information using data visualization techniques
● Design, implement and maintain the software using Data Science
● Take part in software and architectural development activities.
Design, Data Science, Algorithms, Presenting, Science, Testing, Development, Visualization, Software, Internal
2020 - 2020
freelance
Web scraping a US link with potential captcha
Calles Developments USA.
The script will be filling only 5 fields:
Date range: Specific Date Range MM/DD/YYYY to MM/DD/YYYY
Keyword: City (the script will use the attached list with 100 cities which will be updated regularly) or leave it blank if I want to run the script for the whole State
Country: United States (always)
State: Texas (I should be able to choose the state to run the script)
Obituaries: All obituaries (always, as in this case will say “All Texas obituary”)
So, the script should be able to give me the option to run the search by City of by State. If it is run by City, the script will upload the list of cities listed in a spreadsheet.
Keyword for City
So I chose to run the search by City after selecting a date range and the state, then the script will upload the list of cities I want to web scrapped, Once the scripts upload the city list, the script will start web scraping all the deceased people’s information and uploading it to the output file template.
Now while running the script, there could be more than a thousand people death in that specific date range. When that happened, the web page will display the following message: "1000+ Results”, meaning that the search results gave more than 1,000 results. When this happens to a particular city, Then the script will not run the search for that particular city and will make a note on the output file showing that that city gave a "1000+ Results” message.
If the search is runed by State instead of by city and the pages returns a message saying "1000+ Results”, then the script will ask the operator, to reduce the Specific Date Range in order to produce a search result lower than 1,000 results. The script should repeat this step until the search result is lower than 1,000 results.
eywords for Next of Kin
Once the script clicks on “View Obituary” for each person (see attached screenshot), the link will open a new page where there is information about the deceased person. See the attached screenshots that shows all the fields that needs to be web scraped.
When web scraping the highlighted fields in the screenshots, the script will have to upload a list of keywords attached to this request.
This is how the script will use the list of keywords: Once the script is on the page of the deceased’s info page, the script will look for these keywords. Once it finds any of these key words, the scrip will web scrape any information contained between both periods. For example, see the attached screenshot for Melody Ann Link. All the red boxes needs to be web scrapped.
Below, I am showing in red letters some keywords from “Melody Ann”. See how 2 full paragraph needs to be web scraped from period to period:
Melody was preceded in death by her parents, Lillie Bell Wadsworth and Paul Condell Hudson, and her brother, Wesley Paul Hudson. She is survived by her husband, Lloyd Vaughan, of Dallas, Texas, and her son, Wesley Paul Link, of Telluride, Colorado.
In these 2 paragraphs, there are a total of 6 key words from the list sown blow, so the script should web scrape both paragraphs, starting from the 1st period before the name “Melody” and web scrape passing the 2nd period up to the 3rd period just before the word “Colorado”, both words marked on green
Some deceased’s info pages have only very basic information, like name of the deceased person and date of death, or only date of Memorial Service. In this case many columns in the spreadsheet will be empty.
Once the script finishes web scraping all the required data, then it should go back to the search results and keep web scraping the next deceased person’s information.
Names of the deceased person
The script will read the full name of the deceased person and stop as soon as there is a repeated name. Unfortunately, there is no period or some type of indication where the list of names stops when the person had many names, so the script will have to reading the names as soon the next word starts with a small letter or there is a repeated name in the list as shown in one of the screenshots.
Link to the information about the deceased person
Will save each link of each deceased person into the output file so that way I have a way to confirm the source of the information and obtain additional details if I need to.
Date range: Specific Date Range MM/DD/YYYY to MM/DD/YYYY
Keyword: City (the script will use the attached list with 100 cities which will be updated regularly) or leave it blank if I want to run the script for the whole State
Country: United States (always)
State: Texas (I should be able to choose the state to run the script)
Obituaries: All obituaries (always, as in this case will say “All Texas obituary”)
So, the script should be able to give me the option to run the search by City of by State. If it is run by City, the script will upload the list of cities listed in a spreadsheet.
Keyword for City
So I chose to run the search by City after selecting a date range and the state, then the script will upload the list of cities I want to web scrapped, Once the scripts upload the city list, the script will start web scraping all the deceased people’s information and uploading it to the output file template.
Now while running the script, there could be more than a thousand people death in that specific date range. When that happened, the web page will display the following message: "1000+ Results”, meaning that the search results gave more than 1,000 results. When this happens to a particular city, Then the script will not run the search for that particular city and will make a note on the output file showing that that city gave a "1000+ Results” message.
If the search is runed by State instead of by city and the pages returns a message saying "1000+ Results”, then the script will ask the operator, to reduce the Specific Date Range in order to produce a search result lower than 1,000 results. The script should repeat this step until the search result is lower than 1,000 results.
eywords for Next of Kin
Once the script clicks on “View Obituary” for each person (see attached screenshot), the link will open a new page where there is information about the deceased person. See the attached screenshots that shows all the fields that needs to be web scraped.
When web scraping the highlighted fields in the screenshots, the script will have to upload a list of keywords attached to this request.
This is how the script will use the list of keywords: Once the script is on the page of the deceased’s info page, the script will look for these keywords. Once it finds any of these key words, the scrip will web scrape any information contained between both periods. For example, see the attached screenshot for Melody Ann Link. All the red boxes needs to be web scrapped.
Below, I am showing in red letters some keywords from “Melody Ann”. See how 2 full paragraph needs to be web scraped from period to period:
Melody was preceded in death by her parents, Lillie Bell Wadsworth and Paul Condell Hudson, and her brother, Wesley Paul Hudson. She is survived by her husband, Lloyd Vaughan, of Dallas, Texas, and her son, Wesley Paul Link, of Telluride, Colorado.
In these 2 paragraphs, there are a total of 6 key words from the list sown blow, so the script should web scrape both paragraphs, starting from the 1st period before the name “Melody” and web scrape passing the 2nd period up to the 3rd period just before the word “Colorado”, both words marked on green
Some deceased’s info pages have only very basic information, like name of the deceased person and date of death, or only date of Memorial Service. In this case many columns in the spreadsheet will be empty.
Once the script finishes web scraping all the required data, then it should go back to the search results and keep web scraping the next deceased person’s information.
Names of the deceased person
The script will read the full name of the deceased person and stop as soon as there is a repeated name. Unfortunately, there is no period or some type of indication where the list of names stops when the person had many names, so the script will have to reading the names as soon the next word starts with a small letter or there is a repeated name in the list as shown in one of the screenshots.
Link to the information about the deceased person
Will save each link of each deceased person into the output file so that way I have a way to confirm the source of the information and obtain additional details if I need to.
Python, API
2020 - 2020
freelance
BOL Database of Container(TEU) flow in US state/city level by (major) Consignees for last 2 years
Australian Company.
BOL (Bill Of Landing) data for US market with the following details:
YEAR,
MONTH (optional)
Trade Direction (Import/Export)
SCAC (Standard Alpha Carrier Code)
Shipping liner name
nvocc flag
US Port Name
US Port State
US In Bond City Name
Foreign In Bond City (optional)
Foreign In Bond Country
Consignee,
Shipper,
Total TEU
YEAR,
MONTH (optional)
Trade Direction (Import/Export)
SCAC (Standard Alpha Carrier Code)
Shipping liner name
nvocc flag
US Port Name
US Port State
US In Bond City Name
Foreign In Bond City (optional)
Foreign In Bond Country
Consignee,
Shipper,
Total TEU
Python, MongoDB, API, AWS, Html
2019 - 2019
freelance
ParseHub expert to modify existing projects
Caroline Shleifer, Singapore.
ParseHub expert to modify existing ParseHub projects to collect certain information which is not consistently captured in the current projects.
2017 - 2019
job
Data Analyst
D Link Technologies Pvt Ltd.
India
● Created dashboards in Tableau using Excel Files.
● Synthesized current business intelligence data to produce reports and polished
presentations, highlighting findings and recommending changes.
● Created dashboards in Tableau using Excel Files.
● Synthesized current business intelligence data to produce reports and polished
presentations, highlighting findings and recommending changes.
Excel, Business Intelligence, Tableau, Analyst
My education
Middlesex University
Masters, Computer Science
Masters, Computer Science
Jawaharlal Nehru Technological University
Bachelors, Computer Science
Bachelors, Computer Science
ABHIRAM's reviews
ABHIRAM has not received any reviews on Worksome.
Contact ABHIRAM KANNURI
Worksome removes the expensive intermediaries and gives you direct contact with relevant talent.
Create a login and get the opportunity to write to ABHIRAM directly in Worksome.
38000+ qualified freelancers
are ready to help you
Tell us what you need help with
and get specific bids from skilled talent in Denmark