본문 바로가기

Programming/Web Crawling

(14)
sns 크롤링 from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys from bs4 import BeautifulSoup as bs import time url = 'https://www.instagram.com/' driver = wb.Chrome() driver.get(url) # 손으로 로그인 input_search = driver.find_element_by_css_selector('input.XTCLo') input_search.send_keys('찾는키워드') soup = bs(driver.page_source, 'lxml') title_list = soup.select('span._28KuJ + div spa..
이미지 크롤링 from bs4 import BeautifulSoup as bs from selenium import webdriver as wb import time from urllib.request import urlretrieve # 이미지경로를 파일로 저장 from selenium.webdriver.common.keys import Keys url = 'https://search.naver.com/search.naver?where=image&sm=tab_jum&query=%EA%BC%AC%EB%A0%9B' driver = wb.Chrome() driver.get(url) body = driver.find_element_by_tag_name('body') for num in range(50): body.send_..
쇼핑 크롤링 from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys from bs4 import BeautifulSoup as bs import time driver = wb.Chrome() url = 'http://www.11st.co.kr/html/bestSellerMain.html' driver.get(url) titles = [] prices = [] for index in range(5): image_list = driver.find_elements_by_css_selector('div.pub_photo a') time.sleep(2) image_list[index].click() time.sleep(2) sou..
카페 모든 메뉴 가져오기 # 스타벅스 홈페이지 -> menu클릭 from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys from bs4 import BeautifulSoup as bs import time import pandas as pd url = 'https://www.istarbucks.co.kr/index.do' driver = wb.Chrome() driver.get(url) driver.find_element_by_class_name('gnb_nav02').click() #자세히 보기 클릭 driver.find_element_by_class_name('menu_drink_btn01').click() #영양정보 보기 클릭 ..
카페 지점명, 주소, 전화번호 크롤링 from selenium import webdriver as wb from bs4 import BeautifulSoup as bs import time import pandas as pd url = 'http://www.istarbucks.co.kr/store/store_map.do' driver = wb.Chrome() driver.get(url) #지역검색 버튼 클릭 btn_search = driver.find_element_by_class_name('loca_search') #btn_search = driver.find_element_by_xpath('//*[@id="container"]/div/form/fieldset/div/section/article[1]/article/header[2]/h3/..
실습 '사과' 자동검색하기 !pip install selenium import requests as req from bs4 import BeautifulSoup as bs import pandas as pd from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys driver = wb.Chrome() # 웹 페이지 띄워보기 url = 'https://www.google.com/' # 웹 사이트 설정 driver.get(url) input_search = driver.find_element_by_class_name('gLFyf') # 개발자 모드 검색 창 클래스 input_search.send_keys('사과') in..
도시락 크롤링 from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys from bs4 import BeautifulSoup as bs import time import pandas as pd url = 'https://www.hsd.co.kr/menu/menu_list' driver = wb.Chrome() driver.get(url) # 예외처리(try except문) # 더보기 버튼요소를 3번 클릭하기 btn_more = driver.find_element_by_class_name('c_05') try: for index in range(50): btn_more.click() time.sleep(2) #2초동안 멈춤 e..
Selenium 모듈 !pip install selenium from selenium import webdriver as wb from selenium.webdriver.common.keys import Keys #driver.implicitly_wait(5) 최대지연 5초 #웹 브라우저 실행 driver = wb.Chrome() #해당 URL을 브라우저로 실행 url = 'https://www.naver.com' driver.get(url) #검색창 태그(요소) 검색 input_search = driver.find_element_by_id('query') #검색창에 검색어를 입력할 수 있음. input_search.send_keys('날씨') # 검색할 수 있는 2가지 방식 # 1. 버튼 클릭 #검색버튼 태그(요소) 검색 ..