最近用網路爬蟲,
赫然想到是否可以直接用R下載檔案,
發現真的有相關函式,
就是download.file。
你可以給他網址,
並且設定檔案名稱,
即可將檔案下載下來,
配合網路爬蟲抓取多個URL即可自動化下載多個檔案了(?)。
下面有程式碼範例。
When I was using crawler,
I found a interesting function called download.file.
You give it the url of the file you want to download,
and give it a file name.
Then you can download the file with this function.
I think download.file can coordinate with crawler,
and download many files automatically.
Popular Posts
Blog Archive
Categories
R
(28)
data.table
(4)
Python
(3)
Rstudio
(3)
dplyr
(3)
rvest
(3)
網路爬蟲
(3)
Error
(2)
Web Crawler
(2)
grepl
(2)
jupyter
(2)
plyr
(2)
ubuntu
(2)
教學
(2)
.Last.value
(1)
Big Data
(1)
Console
(1)
IEEE程式語言排行
(1)
PuTTY
(1)
Rprofile.site
(1)
Rselenium
(1)
XLConnect
(1)
assign
(1)
bar chart
(1)
cat
(1)
conflict
(1)
coord_flip
(1)
data.frame
(1)
dcast
(1)
download.file
(1)
evalWithTimeout
(1)
excel_sheets
(1)
factor
(1)
file.rename
(1)
fread
(1)
ggplot2
(1)
global variable
(1)
group_by
(1)
gsub
(1)
invalid multibyte character
(1)
jiebaR
(1)
join
(1)
jupyter_contrib_nbextensions
(1)
jupyterthemes
(1)
loading
(1)
melt
(1)
merge
(1)
mutate
(1)
numeric
(1)
print
(1)
rbind
(1)
read.csv
(1)
read_csv
(1)
read_excel
(1)
readr
(1)
readxl
(1)
scientific notation
(1)
scipen
(1)
separate_rows
(1)
setDF
(1)
setDT
(1)
sqldf
(1)
static IP address
(1)
str_count
(1)
stringr
(1)
table
(1)
tidyr
(1)
timeout
(1)
trim
(1)
txtProgressBar
(1)
unique
(1)
zip
(1)
人力銀行
(1)
參考資源
(1)
技能
(1)
文字探勘
(1)
橫條圖
(1)
玩玩小數據
(1)
結巴分詞
(1)
能力
(1)
資料分析
(1)
資料分析師
(1)
長條圖
(1)
沒有留言:
張貼留言