-
-
Notifications
You must be signed in to change notification settings - Fork 702
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cannot download with read.csv() #37
Comments
For the scientific purpose, I will offer another link for you to reach out for the CSV raw content soon. Please wait for about 30 minutes, I will try to fix it as soon as possible. |
Hello, I have fixed this issue. If the aforementioned methods are too complicated for you, you can just switch the URL to https://lab.isaaclin.cn/csv/DXYArea.csv, and everything will just work. All the four CSV files can be accessed from this link by changing the filename. However, these CSV files are stored on my server, so it will be a little bit slower than downloading it from GitHub. Furthermore, every time you run this line, your code will actually download this file, and there will be a relatively heave traffic on the server-side. Therefore, please try to load it as fewer times as possible and backup at first, for example: # Just load once
DTdxy_backup <- read.csv("https://lab.isaaclin.cn/csv/DXYArea.csv", header = TRUE, stringsAsFactors = FALSE)
# Do research on DTdxy
DTdxy <- DTdxy_backup
# If you messed up with DTdxy, reload it from DTdxy_backup
DTdxy <- DTdxy_backup Hope you can enjoy your research. |
Yes, the new address worked well for me; However, the new dataset lack data before 2020-1-24; The earliest time should be on 2020-1-11. I use the following code and found data before that date is missing: DTdxy_backup <- read.csv("https://lab.isaaclin.cn/csv/DXYArea.csv",
header = TRUE, stringsAsFactors = FALSE)
> DTWuhan <- DTdxy_backup[DTdxy_backup$cityEnglishName=="Wuhan",]
> DTWuhan %>%
+ mutate(Date = as.Date(updateTime)) %>%
+ group_by(Date) %>%
+ filter(updateTime == max(updateTime)) %>%
+ select(cityEnglishName,city_confirmedCount,city_suspectedCount,
+ city_curedCount,city_deadCount,Date) %>% tail()
# A tibble: 6 x 6
# Groups: Date [6]
cityEnglishName city_confirmedCount city_suspectedCount city_curedCount city_deadCount Date
<chr> <int> <int> <int> <int> <date>
1 Wuhan 1905 0 54 104 2020-01-29
2 Wuhan 1590 0 47 85 2020-01-28
3 Wuhan 698 0 42 63 2020-01-27
4 Wuhan 618 0 40 45 2020-01-26
5 Wuhan 572 0 32 38 2020-01-25
6 Wuhan 495 0 31 23 2020-01-24 Can you help me to retrieve the data before 2020-1-24? |
Currently, only province-level data after January 22 and city-level data after January 24 are obtained and available by the crawler. Those data are published by Ding Xiang Yuan and supposed to be trust-worthy. However, most of the data from January 11 to January 24 are missing and do not have a reliable source. If you are urgent, please try to find some reliable data sources and let me know. Now I have 2 data sources but the values are slightly different from each other, so I do not add them into the database and this data warehouse yet. Now, I am the only one maintaining this project, but I'm not a professional in this field, so I don't know many data sources in this field. Currently, I have to cross-validate the data sources, and it will take a relatively long period of time. If you can find some reliable data sources, please contribute to this project. The data will be collected and added into the database very soon once there are reliable data sources. Open source projects are not projects maintained and dedicated by a single person. |
Thank you for your clarification. I will try to contribute this project if I can find some reliable source. |
Thank you so much! Currently, this issue is widely discussed in BlankerL/DXY-COVID-19-Crawler#3 and #26, and 2 data sources I mentioned above are in these issues. |
It raise an error:
With a warning message:
And when I copy the https://raw.githubusercontent.com/BlankerL/DXY-COVID-19-Data/master/csv/DXYArea.csv to browser, the connection cannot be opened either.
The text was updated successfully, but these errors were encountered: