ªð¦^¦Cªí ¤W¤@¥DÃD µo©«

[­ì³Ð] python¤W¥«Âd¤T¤jªk¤H¶R½æ¶W¤é³ø¸ê®Æ¤U¸ü

¥»©«³Ì«á¥Ñ lpk187 ©ó 2016-8-31 09:05 ½s¿è

¦^´_ 21# c_c_lai

jupyter©MIDLE°õ¦æ¦³ÂI¤£¦P¡Ajupyter¬O¤@¬q¤@¬q°õ¦æªº¡A
°õ¦æ¬Y¬q¥²¶·Ä~©Ó¤W¤@¦¸ªº°õ¦æ¡A§_«h·|¿ù»~
¦Ó°õ¦æªºµ²ªG¡A«h¬O¦b¬Y¬qªº¤U­±¡AShift Enter ¬O¹ïªº

   

TOP

±ÀÂ˨ϥΠAnaconda ªº Python ª©¥», ¦]¬°³o­Óª©¥»À°§A§â³\¦h®M¥ó (¤×¨ä¬O¬ì¾Ç­pºâ¡B¸ê®Æ¤ÀªR¬ÛÃöªº) ³£¸Ë¦n¤F¡C

    https://www.continuum.io/downloads

TOP

¦^´_ 38# c_c_lai


    soup = BeautifulSoup(res.text,"html.parser")

TOP

¦^´_ 41# c_c_lai

    ªí®æ¥i¥H¥Î pandas ¥h§ì¸Ô²Óªº«ü¥O¥i°Ñ·Ó http://pandas.pydata.org/pandas- ... tml#io-excel-writer
  1. import pandas as pd
  2. import io
  3. url = 'http://www.taifex.com.tw/chinese/3/7_12_3.asp'
  4. dfs = pd.read_html(url, index_col=0)
  5. data=dfs[3][0:][1:]
  6. data.to_excel('test.xlsx')
½Æ»s¥N½X
¤£¹L§ì¦^¨Óªºªí®æ¦³ÂI¿ù¶Ã¡A§Ú¤]¬O­è¶}©l¾Ç¡A§O¤¶·N¡I
¤j®a¤¬¬Û¬ã¨s¡I   :$

TOP

¦^´_ 41# c_c_lai

html¤ñ¸û¦h°ÝÃD¡A ·í§A¦³html5lib / LXML¸ÑªR¾¹ªº°ÝÃD
§A¥i¥H¤U¸ü¥¦
pip install html5lib
easy_install lxml

TOP

¥»©«³Ì«á¥Ñ lpk187 ©ó 2016-9-8 09:16 ½s¿è

¦^´_ 47# c_c_lai

¦w¸Ë¼Ò¶ô³Ì¦n¦bDOS¤¤¦w¸Ë ¡A¨ä¹ê¦b¦w¸ËPython®É¤j¦h¤w¼g¤JÀô¹ÒÅܼơA¦ý¬O³Ì¦nÁÙ¬O¦b¨äª©¥»¤UªºScripts¸ê®Æ§¨¦w¸Ë¤ñ¸û¦n
¨Ò¦pAnaconda3ª©¥»¤U¡G¥ý¶i¤Jcmd ¶i¤JDOS «á¥´¤Jcd  C:\Anaconda3\Scripts    <<==¨Ì§A¹ê»Ú¸ê®Æ§¨¬°·Ç
¦w¸Ë«e¡A¥ý¤É¯Åpip  »yªk¡G"pip install --upgrade pip"
pandas¡Gpip install pandas
¦w¸Ëpandasªº¦P®É¥¦¦n¹³¤]·|À°§A¦w¸Ë

html5lib¤]¬Opip¡Gpip install html5lib ¡A­Y¤w¸g¦³¦w¸Ë¤F¡A¨º´N¤É¯Å¥¦ pip install -U html5lib -U¬O¤É¯Åªº·N«ä U ¥²¶·¤j¼g

¦Ü©ó lxml ¤j¦h¤w¸g¦w¸Ë¡A¦ýª©¥»®e©ö¥X¿ù¡A©Ò¥H­n¥ý¤Ï¦w¸Ë  conda remove lxml
lxml ¥Îpip¤£¦n¦w¸Ë¡A©Ò¥H§Ú¬O¥Î easy_install ¦w¸Ëªº  easy_install lxml

¨ä¥Lª©¥»¤]¬O¤@¼Ë»Ý­n¨ì¨ä¥Ø¿ý(Scripts)¦w¸Ë¤ñ¸û§´·í

TOP

¦^´_ 49# c_c_lai


    ¬Oªº¤@¼Ë¶i¤J¨ìpython3.5¸ê®Æ§¨ªºScripts¸ê®Æ§¨¶i¦æ ¦w¸Ë  
­Y¬O¦b¨ä¥Lªº¸ê®Æ§¨©Î®Ú¥Ø¿ý¤U¦w¸Ë¡A¤ñ¸û·|¿ù¶Ã¡A¨ä¦]¬O¦w¸Ë¦h­Óª©¥»ªº¶}«Y¡I
¦³®Épip¤£¯à¤U¸ü¦w¸Ë¥i¥H¹Á¸Õ¨ä¥L¦w¸Ë¤èªk ¨Ò¦p¥Î easy_install ¦w¸Ë©ÎªÌ°Ñ¦Ò ¨ä¼Ò¶ôªº©xºô»¡©ú

TOP

¦^´_ 50# c_c_lai


    ©Ò²£¥ÍªºÀÉ®×·|¦b³o­Ó©ÒÀx¦sµ{¦¡½X¤Uªº¸ê®Æ§¨¥X²{
¥Hjupyter ¨Ó»¡¤@¯ë³£¦b"¤å¥ó"¸ê®Æ¤¤

TOP

¦^´_ 84# c_c_lai

§Ú©M§A¤@¼Ë¤]¥u¬Oªì¾ÇPython ¡A¹ï©óPython¤]©|¦b¬ã¨s¤§¤¤¡A
¬Q¤Ñ¬Ý¨ì§A©Mzyzzyva¤j¤j¬ã¨s³o½g­Ë¤]¬O¤Þµo§Ú«Ü¤jªº¿³½ì¡I
¨ä¹ê§Ú¹ï pandas »{ÃѤ]¤£²`¡A¤]¶È¬O­è¦n¬Ý¨ì pandas ¥i¥H»s§@ªí®æ °Ñ¦Òºô§}
¨ä¥Lªº¥i¤£¤ÓÀ´ªº¡A HTML ¥N½X¡C¹ï§Ú¨Ó»¡¡A¥i¬O³B¦b§¹¥þ¤£À´°Ú¡I­ü¡I
¹ïBeautiful Soup 4¡A§c¡I¤]¬O¬Ý±oªáªáªº ¡A°Ñ¦Òºô§}
¤@°_¥[ªo§a¡I

TOP

¥»©«³Ì«á¥Ñ lpk187 ©ó 2016-9-12 23:46 ½s¿è

¦^´_ 103# koshi0413

¬Ý¨Ó¼ö°J©ópython ªº¦P¦n¦b³o°Q½×°Ï¡A¤]ÆZ¦h¤Hªº
python ªºio¥i¥H¿é¥X«Ü¦h®æ¦¡¡Acsv .xlsx¬Æ¦Üsql³£¦æ

·íC¤j¶}©l°Q½×³oºô§} http://www.twse.com.tw/ch/trading/exchange/MI_MARGN/MI_MARGN.php
´N¦b·Q¡Aªí®æÁÙ¬O¥æµ¹ªí®æ±M®a pandas ¨Ó³B²z¡A©ó¬O­É¤F§A¤j³¡¥÷¥N½X¨Ó°µ¬Ý¬Ý¡A®ÄªGÁÙ¤£¿ù
  1. import requests
  2. from bs4 import BeautifulSoup
  3. import pandas as pd
  4. import io

  5. url = 'http://www.twse.com.tw/ch/trading/exchange/MI_MARGN/MI_MARGN.php'
  6. headers = {"User-Agent":"Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/49.0.2623.112 Safari/537.36"}
  7. url1 = "http://www.twse.com.tw/ch/trading/exchange/MI_MARGN/MI_MARGN.php"
  8. payload = {"download":'',
  9.             "qdate":'105/09/10',
  10.             "selectType":"ALL"}
  11. res = requests.post(url1, headers=headers, data=payload)
  12. tbl=pd.read_html(res.text)

  13. deta=tbl[1]
  14. deta.columns = ['ªÑ²¼¥N¸¹','ªÑ²¼¦WºÙ','¶R¶i','½æ¥X','²{ª÷ÀvÁÙ','«e¤é¾lÃB','¤µ¤é¾lÃB','­­ÃB','¶R¶i','½æ¥X','²{¨éÀvÁÙ','«e¤é¾lÃB','¤µ¤é¾lÃB','­­ÃB','¸ê¨é¤¬©è','µù°O']
  15. deta.to_csv('test1.csv')
½Æ»s¥N½X

TOP

        ÀR«ä¦Û¦b : ¡i®É¶¡µLªk¾B¾×¡j©È®É¶¡®ø³u¡Aªá¤F³\¦h¤ß¦å¡A·QºÉ¦U¦¡¤èªk­n¾B¾×®É¶¡¡Aµ²ªG¬O¡G®ö¶O¤F§ó¦h®É¶¡¡A¥B¤@µL©Ò¦¨¡I
ªð¦^¦Cªí ¤W¤@¥DÃD