AssistScraper - program for /r/nba to use to find list of all players a player assisted and how many assists each player recieved

Overview

AssistScraper

program for /r/nba to use to find list of all players a player assisted and how many assists each player recieved

################################

Readme.txt

how to use this program for dummies

aimed at people who don't know what they're doing (/r/nba)

I tell you to install a lot of things here, and I know that sounds sketchy but it really is safe to do

trust me I'm some guy on the internet

################################

Step 1- install python

This is what will run the program

google install python

click on the python.org link

click on whatever is there for "download the latest version for windows"

Step 2 - download pycharm

This is what I used to write/debug the program, so I recommend using it to get that nice output

it'll also make it easier for dummies like us to install the correct packages

google install pycharm

click on the jetbrains.com link, then click get pycharm in top right

download the community link

Step 3 - Open the program

This is dumb, but probably the easiest way to do this

Open pycharm, start a new project, name it whatever

delete all the text in the default file that opens

literally copy and paste all of main.py from my github repo to that page

some of the lines at the top will probably be red, meaning you dont have the packages I imported

hover over them and a red lightbulb should pop up. First thing on the list of options it gives you should be "Install ..."

click that

if that doesn't work for some reason, click "Python Packages" on the bottom

type the packages that didn't work in and install them

Step 4 - Have fun!

scroll down to the bottom where it says "your code here" and run a call of getAssists() with your preferred player's url

You might also like...
A training task for web scraping using python multithreading and a real-time-updated list of available proxy servers.

Parallel web scraping The project is a training task for web scraping using python multithreading and a real-time-updated list of available proxy serv

A Very simple free proxy list scraper.
A Very simple free proxy list scraper.

Scrappp A Very simple free proxy list scraper, made in python The tool scrape proxy from diffrent sites and api's. Screenshots About the script !!! RE

A list of Python Bots used to extract data from several websites

A list of Python Bots used to extract data from several websites. Data extraction is for products on e-commerce (ecommerce) websites. Data fetched i

A command-line program to download media, like and unlike posts, and more from creators on OnlyFans.
A command-line program to download media, like and unlike posts, and more from creators on OnlyFans.

onlyfans-scraper A command-line program to download media, like and unlike posts, and more from creators on OnlyFans. Installation You can install thi

This program scrapes information and images for movies and TV shows.

Media-WebScraper This program scrapes information and images for movies and TV shows. Summary For more information on the program, read the WebScrape_

Web-scraping - Program that scrapes a website for a collection of quotes, picks one at random and displays it
Web-scraping - Program that scrapes a website for a collection of quotes, picks one at random and displays it

web-scraping Program that scrapes a website for a collection of quotes, picks on

a Scrapy spider that utilizes Postgres as a DB, Squid as a proxy server, Redis for de-duplication and Splash to render JavaScript. All in a microservices architecture utilizing Docker and Docker Compose

This is George's Scraping Project To get started cd into the theZoo file and run: chmod +x script.sh then: ./script.sh This will spin up a Postgres co

An application that on a given url, crowls a web page and gets all words,  sorts and counts them.
An application that on a given url, crowls a web page and gets all words, sorts and counts them.

Web-Scrapping-1 An application that on a given url, crowls a web page and gets all words, sorts and counts them. Installation Using the package manage

Command line program to download documents from web portals.

command line document download made easy Highlights list available documents in json format or download them filter documents using string matching re

Comments
  • Assists for a particular season?

    Assists for a particular season?

    Hi Akarabat There is no issue. Just thought I would ask if there is a way to only get the assists data for a particular season? I do not have any coding experience. Just like to check on the data. thanks

    Ol

    opened by LePetitNewbie 5
Owner
null
WebScraper - A script that prints out a list of all EXTERNAL references in the HTML response to an HTTP/S request

Project A: WebScraper A script that prints out a list of all EXTERNAL references

null 2 Apr 26, 2022
Scrapping the data from each page of biocides listed on the BAUA website into a csv file

Scrapping the data from each page of biocides listed on the BAUA website into a csv file

Eric DE MARIA 1 Nov 30, 2021
VG-Scraper is a python program using the module called BeautifulSoup which allows anyone to scrape something off an website. This program lets you put in a number trough an input and a number is 1 news article.

VG-Scraper VG-Scraper is a convinient program where you can find all the news articles instead of finding one yourself. Installing [Linux] Open a term

null 3 Feb 13, 2022
This program will help you to properly scrape all data from a specific website

This program will help you to properly scrape all data from a specific website

MD. MINHAZ 0 May 15, 2022
Free-Game-Scraper is a useful script that allows you to track down free games and DLCs on many platforms.

Game Scraper Free-Game-Scraper is a useful script that allows you to track down free games and DLCs on many platforms. Join the discord About The Proj

KursK 2 Mar 28, 2022
This is a simple website crawler which asks for a website link from the user to crawl and find specific data from the given website address.

This is a simple website crawler which asks for a website link from the user to crawl and find specific data from the given website address.

Faisal Ahmed 1 Jan 10, 2022
Find papers by keywords and venues. Then download it automatically

paper finder Find papers by keywords and venues. Then download it automatically. How to use this? Search CLI python search.py -k "knowledge tracing,kn

Jiahao Chen (TabChen) 2 Dec 15, 2022
PS5 bot to find a console in france for chrismas 🎄🎅🏻 NOT FOR SCALPERS

Une PS5 pour Noël Python + Chrome --headless = une PS5 pour noël MacOS Installer chrome Tweaker le .yaml pour la listes sites a scrap et les criteres

Olivier Giniaux 3 Feb 13, 2022
A web Scraper for CSrankings.com that scrapes University and Faculty list for a particular country

A web Scraper for CSrankings.com that scrapes University and Faculty list for a particular country To run the file: Open terminal

null 2 Jun 6, 2022
Instagram_scrapper - This project allow you to scrape the list of followers, following or both from a public Instagram account, and create a csv or excel file easily.

Instagram_scrapper This project allow you to scrape the list of followers, following or both from a public Instagram account, and create a csv or exce

Lakhdar Belkharroubi 5 Oct 17, 2022