Skip to content

Make fixes #193

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 11 commits into
base: main
Choose a base branch
from
Open

Make fixes #193

wants to merge 11 commits into from

Conversation

Karim-Mane
Copy link
Member

No description provided.

@Karim-Mane Karim-Mane requested a review from avallecam June 16, 2025 10:56
Copy link

github-actions bot commented Jun 16, 2025

🆗 Pre-flight checks passed 😃

This pull request has been checked and contains no modified workflow files or spoofing.

Results of any additional workflows will appear here when they are done.

github-actions bot pushed a commit that referenced this pull request Jun 16, 2025
github-actions bot pushed a commit that referenced this pull request Jun 16, 2025
Copy link
Contributor

@Degoot-AM Degoot-AM left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @Karim-Mane for the suggestions, which improve the overall readability of this tutorials.

github-actions bot pushed a commit that referenced this pull request Jun 23, 2025
Comment on lines +30 to +32
In the process of analyzing outbreak data, it's essential to ensure that the dataset is clean, curated, standardized, and validated. This will increase the accuracy (i.e. you are analysing what you think you are analysing) and reproducibility (i.e. if someone wants to go back and repeat your analysis steps with your code, you can be confident they will get the same results) of the
downstream analyses. This episode focuses on cleaning epidemics and outbreaks data using the [{cleanepi}](https://epiverse-trace.github.io/cleanepi/) package.
To demonstrate the functionalities of this package, we'll apply them on a simulated dataset of Ebola cases.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
In the process of analyzing outbreak data, it's essential to ensure that the dataset is clean, curated, standardized, and validated. This will increase the accuracy (i.e. you are analysing what you think you are analysing) and reproducibility (i.e. if someone wants to go back and repeat your analysis steps with your code, you can be confident they will get the same results) of the
downstream analyses. This episode focuses on cleaning epidemics and outbreaks data using the [{cleanepi}](https://epiverse-trace.github.io/cleanepi/) package.
To demonstrate the functionalities of this package, we'll apply them on a simulated dataset of Ebola cases.
In the process of analyzing outbreak data, it's essential to ensure that the dataset is clean, curated, standardized, and validated. This will increase the accuracy (i.e. you are analysing what you think you are analysing) and reproducibility (i.e. if someone wants to go back and repeat your analysis steps with your code, you can be confident they will get the same results) of the downstream analyses. This episode focuses on cleaning epidemics and outbreaks data using the [{cleanepi}](https://epiverse-trace.github.io/cleanepi/) package.
To demonstrate the functionalities of this package, we'll apply them on a simulated dataset of Ebola cases.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

text alignment

Comment on lines +30 to +32
In the process of analyzing outbreak data, it's essential to ensure that the dataset is clean, curated, standardized, and validated. This will increase the accuracy (i.e. you are analysing what you think you are analysing) and reproducibility (i.e. if someone wants to go back and repeat your analysis steps with your code, you can be confident they will get the same results) of the
downstream analyses. This episode focuses on cleaning epidemics and outbreaks data using the [{cleanepi}](https://epiverse-trace.github.io/cleanepi/) package.
To demonstrate the functionalities of this package, we'll apply them on a simulated dataset of Ebola cases.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

text alignment

Comment on lines +72 to +73
The `::` in R lets you access functions or objects from a specific package without attaching the entire package to the search path. It offers several important
advantages including the followings:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
The `::` in R lets you access functions or objects from a specific package without attaching the entire package to the search path. It offers several important
advantages including the followings:
The `::` in R lets you access functions or objects from a specific package without attaching the entire package to the search path. It offers several important advantages including the followings:

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

text alignment

Comment on lines +214 to +215
You can get the number and location of the duplicated rows that where found. Run `cleanepi::print_report()`, wait for the report to open in your browser, and
find the "Duplicates" tab.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
You can get the number and location of the duplicated rows that where found. Run `cleanepi::print_report()`, wait for the report to open in your browser, and
find the "Duplicates" tab.
You can get the number and location of the duplicated rows that where found. Run `cleanepi::print_report()`, wait for the report to open in your browser, and find the "Duplicates" tab.

Comment on lines +294 to +295
Each entry in the dataset represents a subject (e.g. a disease case or study participant) and should be distinguishable by a specific ID formatted in a
particular way. These IDs can contain numbers falling within a specific range, a prefix and/or suffix, and might be written such that they contain a specific number of characters. The `{cleanepi}` package offers the function `check_subject_ids()` designed precisely for this task as shown in the below code chunk. This function checks whether the IDs are unique and meet the required criteria specified by the user.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Each entry in the dataset represents a subject (e.g. a disease case or study participant) and should be distinguishable by a specific ID formatted in a
particular way. These IDs can contain numbers falling within a specific range, a prefix and/or suffix, and might be written such that they contain a specific number of characters. The `{cleanepi}` package offers the function `check_subject_ids()` designed precisely for this task as shown in the below code chunk. This function checks whether the IDs are unique and meet the required criteria specified by the user.
Each entry in the dataset represents a subject (e.g. a disease case or study participant) and should be distinguishable by a specific ID formatted in a particular way. These IDs can contain numbers falling within a specific range, a prefix and/or suffix, and might be written such that they contain a specific number of characters. The `{cleanepi}` package offers the function `check_subject_ids()` designed precisely for this task as shown in the below code chunk. This function checks whether the IDs are unique and meet the required criteria specified by the user.

@@ -180,7 +184,7 @@ db_connection <- DBI::dbConnect(

A real-life connection to an external SQLite database would look like this:

```r
```{r}
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
```{r}
```{r, eval=FALSE}
Suggested change
```{r}
```{r}

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do not run this code because it not excuetable

github-actions bot pushed a commit that referenced this pull request Jun 26, 2025
@avallecam
Copy link
Member

avallecam commented Jul 3, 2025

Hi @Karim-Mane, let me know when this is ready to review. I'll be able to review and assess merging tomorrow, ready for the start of next week's training.

I can deal with solving some of the remining gh-action issues

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants