Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reviewed general punctuation and fixed some broken links #251

Merged
merged 58 commits into from
Jul 17, 2024
Merged
Show file tree
Hide file tree
Changes from 10 commits
Commits
Show all changes
58 commits
Select commit Hold shift + click to select a range
227e279
Punctuation updated and broken image link restored
sergiopaniego Apr 12, 2024
87a9456
Review punctuation and fixed some broken links
sergiopaniego Apr 15, 2024
a7bb7db
Merge remote-tracking branch 'upstream/main' into main
sergiopaniego Apr 23, 2024
aee83e5
Restored HTTP image link
sergiopaniego Apr 23, 2024
5b0b687
Merge branch 'main' of https://github.com/johko/computer-vision-cours…
sergiopaniego Apr 24, 2024
fdc9f2f
Merge remote-tracking branch 'upstream/main'
sergiopaniego Apr 26, 2024
7efc70c
Update chapters/en/unit5/generative-models/gans-vaes/stylegan.mdx
sergiopaniego Apr 26, 2024
e84231b
Update chapters/en/unit5/generative-models/gans-vaes/stylegan.mdx
sergiopaniego Apr 26, 2024
5a071d4
Updated suggestions and broken link
sergiopaniego Apr 26, 2024
63c5921
Merge remote-tracking branch 'upstream/main'
sergiopaniego Apr 26, 2024
9f775d5
Merge remote-tracking branch 'remote_repo/main' into main
sergiopaniego May 15, 2024
018cef9
Update chapters/en/unit0/welcome/welcome.mdx
sergiopaniego May 15, 2024
2e33328
Update chapters/en/unit0/welcome/welcome.mdx
sergiopaniego May 15, 2024
7516792
Update chapters/en/unit0/welcome/welcome.mdx
sergiopaniego May 15, 2024
b6be1d8
Update chapters/en/unit0/welcome/welcome.mdx
sergiopaniego May 15, 2024
3953ca7
Update chapters/en/unit0/welcome/welcome.mdx
sergiopaniego May 15, 2024
efd687e
Updated punctuation
sergiopaniego May 15, 2024
17f63fe
Update chapters/en/unit10/blenderProc.mdx
sergiopaniego May 15, 2024
90d441e
Update chapters/en/unit10/blenderProc.mdx
sergiopaniego May 15, 2024
50a4fec
Update chapters/en/unit10/blenderProc.mdx
sergiopaniego May 15, 2024
3857b6b
Update chapters/en/unit10/point_clouds.mdx
sergiopaniego May 15, 2024
7082512
Update chapters/en/unit10/synthetic_datasets.mdx
sergiopaniego May 15, 2024
63f11a9
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
b531fd3
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
9e34648
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
5ba57b4
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
6a17a60
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
d987aeb
Update chapters/en/unit2/cnns/convnext.mdx
sergiopaniego May 15, 2024
93cc5dc
Update chapters/en/unit3/vision-transformers/detr.mdx
sergiopaniego May 15, 2024
ccbd8ca
Update chapters/en/unit3/vision-transformers/vision-transformer-for-o…
sergiopaniego May 15, 2024
16c7f5e
Update chapters/en/unit4/multimodal-models/clip-and-relatives/clip.mdx
sergiopaniego May 15, 2024
011d714
Update chapters/en/unit4/multimodal-models/clip-and-relatives/clip.mdx
sergiopaniego May 15, 2024
fb6a7c6
Update chapters/en/unit5/generative-models/diffusion-models/introduct…
sergiopaniego May 15, 2024
da41433
Update chapters/en/unit5/generative-models/diffusion-models/introduct…
sergiopaniego May 15, 2024
a23ecc0
Update chapters/en/unit5/generative-models/diffusion-models/introduct…
sergiopaniego May 15, 2024
56ad356
Update chapters/en/unit5/generative-models/diffusion-models/introduct…
sergiopaniego May 15, 2024
c8e9a55
Update chapters/en/unit5/generative-models/diffusion-models/stable-di…
sergiopaniego May 15, 2024
279c808
Update chapters/en/unit5/generative-models/diffusion-models/stable-di…
sergiopaniego May 15, 2024
9e4eb7b
Update chapters/en/unit5/generative-models/introduction/introduction.mdx
sergiopaniego May 15, 2024
034d3fb
Update chapters/en/unit5/generative-models/introduction/introduction.mdx
sergiopaniego May 15, 2024
2937fbc
Update chapters/en/unit5/generative-models/introduction/introduction.mdx
sergiopaniego May 15, 2024
df87c30
Update chapters/en/unit5/generative-models/practical-applications/eth…
sergiopaniego May 15, 2024
c134cb8
Update chapters/en/unit5/generative-models/practical-applications/eth…
sergiopaniego May 15, 2024
8d927af
Update chapters/en/unit5/generative-models/practical-applications/eth…
sergiopaniego May 15, 2024
16a17a0
Update chapters/en/unit8/3d_measurements_stereo_vision.mdx
sergiopaniego May 15, 2024
f74dce2
Update chapters/en/unit8/3d_measurements_stereo_vision.mdx
sergiopaniego May 15, 2024
99450b6
Update chapters/en/unit8/3d_measurements_stereo_vision.mdx
sergiopaniego May 15, 2024
1c55514
Update chapters/en/unit8/introduction/brief_history.mdx
sergiopaniego May 15, 2024
39299fa
Update chapters/en/unit8/introduction/brief_history.mdx
sergiopaniego May 15, 2024
7e8bbc0
Modified pip install spacing
sergiopaniego May 15, 2024
c122f3e
Merge branch 'main' of https://github.com/sergiopaniego/computer-visi…
sergiopaniego May 15, 2024
71233b6
Extended spacing update of pip install instructions across units
sergiopaniego May 15, 2024
6107b34
Removed punctuations in table following standards
sergiopaniego May 15, 2024
c652dfb
Removed unneeded new lines
sergiopaniego May 15, 2024
1c20a24
Updated some broken links pointing to the course
sergiopaniego May 15, 2024
a39c49d
Added new line to separate sentences
sergiopaniego May 16, 2024
3f73731
Removed unneeded puntuaction
sergiopaniego May 16, 2024
1c69233
Removed unneeded punctuation
sergiopaniego May 16, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
15 changes: 8 additions & 7 deletions chapters/en/unit0/welcome/welcome.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -12,29 +12,30 @@ On this page, you can find how to join the learners community, make a submission

To obtain your certification for completing the course, complete the following assignments:

1. Training/fine-tuning a Model
2. Building an application and hosting it on Hugging Face Spaces
1. Training/fine-tuning a Model.
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
2. Building an application and hosting it on Hugging Face Spaces.
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved

### Training/fine-tuning a Model

There are notebooks under the Notebooks/Vision Transformers section. As of now, we have notebooks for object detection, image segmentation, and image classification. You can either train a model on a dataset that exists on 🤗 Hub or upload a dataset to a dataset repository and train a model on that.

The model repository needs to have the following:

1. A properly filled model card, you can check out [here for more information](https://huggingface.co/docs/hub/en/model-cards)

1. A properly filled model card, you can check out [here for more information](https://huggingface.co/docs/hub/en/model-cards).
2. If you trained a model with transformers and pushed it to Hub, the model card will be generated. In that case, edit the card and fill in more details.
3. Add the dataset’s ID to the model card to link the model repository to the dataset repository.

### Creating a Space

In this assignment section, you'll be building a Gradio-based application for your computer vision model and sharing it on 🤗 Spaces. Learn more about these tasks using the following resources:

- [Getting started with Gradio](https://huggingface.co/learn/nlp-course/chapter9/1?fw=pt#introduction-to-gradio)
- [How to share your application on 🤗 Spaces](https://huggingface.co/learn/nlp-course/chapter9/4?fw=pt)
- [Getting started with Gradio](https://huggingface.co/learn/nlp-course/chapter9/1?fw=pt#introduction-to-gradio).
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
- [How to share your application on 🤗 Spaces](https://huggingface.co/learn/nlp-course/chapter9/4?fw=pt).
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved

## Certification 🥇

Once you've finished the assignments — Training/fine-tuning a Model and Creating a Space — please complete the [form](https://forms.gle/isiVSw59oiiHP6pN9) with your name, email, and links to your model and Space repositories to receive your certificate
Once you've finished the assignments — Training/fine-tuning a Model and Creating a Space — please complete the [form](https://forms.gle/isiVSw59oiiHP6pN9) with your name, email, and links to your model and Space repositories to receive your certificate.

## Join the community!

Expand All @@ -52,7 +53,7 @@ As a computer vision course learner, you may find the following set of channels

- `#computer-vision`: a catch-all channel for everything related to computer vision.
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
- `#cv-study-group`: a place to exchange ideas, ask questions about specific posts and start discussions.
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
- `#3d`: a channel to discuss aspects of computer vision specific to 3D computer vision
- `#3d`: a channel to discuss aspects of computer vision specific to 3D computer vision.
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved

If you are interested in generative AI, we also invite you to join all channels related to the Diffusion Models: #core-announcements, #discussions, #dev-discussions, and #diff-i-made-this.

Expand Down
8 changes: 4 additions & 4 deletions chapters/en/unit1/feature-extraction/feature-matching.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Imagine you have a giant box of puzzle pieces, and you're trying to find a speci

Now that we have an intuitive idea of how brute-force matches are found, let's dive into the algorithms. We are going to use the descriptors that we learned about in the previous chapter to find the matching features in two images.

First install and load libraries
First install and load libraries.

```bash
!pip install opencv-python
Expand Down Expand Up @@ -137,13 +137,13 @@ We also create a dictionary to specify the maximum leafs to visit as follows.
search_params = dict(checks=50)
```

Initiate SIFT detector
Initiate SIFT detector.

```python
sift = cv.SIFT_create()
```

Find the keypoints and descriptors with SIFT
Find the keypoints and descriptors with SIFT.

```python
kp1, des1 = sift.detectAndCompute(img1, None)
Expand Down Expand Up @@ -259,7 +259,7 @@ Fm, inliers = cv2.findFundamentalMat(mkpts0, mkpts1, cv2.USAC_MAGSAC, 0.5, 0.999
inliers = inliers > 0
```

Finally, we can visualize the matches
Finally, we can visualize the matches.

```python
draw_LAF_matches(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ In digital image processing, operations on images are diverse and can be categor
- Statistical
- Geometrical
- Mathematical
- Transform operations.
- Transform operations

Each category encompasses different techniques, such as morphological operations under logical operations or fourier transforms and principal component analysis (PCA) under transforms. In this context, we refer to morphology as the group of operations that use structuring elements to generate images of the same size by looking into the values of the pixel neighborhood. Understanding the distinction between element-wise and matrix operations is important in image manipulation. Element-wise operations, such as raising an image to a power or dividing it by another image, involve processing each pixel individually. This pixel-based approach contrasts with matrix operations, which utilize matrix theory for image manipulation. Having said that, you can do whatever you want with images, as they are matrices containing numbers!

Expand Down
2 changes: 1 addition & 1 deletion chapters/en/unit1/image_and_imaging/imaging.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ The core of digital image formation is the function \\(f(x,y)\\), which is deter
</div>

In transmission-based imaging, such as X-rays, transmissivity takes the place of reflectivity. The digital representation of an image is essentially a matrix or array of numerical values, each corresponding to a pixel. The process of transforming continuous image data into a digital format is twofold:
- Sampling, which digitizes the coordinate values
- Sampling, which digitizes the coordinate values.
- Quantization, which converts amplitude values into discrete quantities.

The resolution and quality of a digital image significantly depend on the following:
Expand Down
14 changes: 9 additions & 5 deletions chapters/en/unit10/blenderProc.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -103,18 +103,22 @@ You can install BlenderProc via pip:

Alternately, you can clone the official [BlenderProc repository](https://github.com/DLR-RM/BlenderProc) from GitHub using Git:

`git clone https://github.com/DLR-RM/BlenderProc`
```bash
git clone https://github.com/DLR-RM/BlenderProc
```

BlenderProc must be run inside the blender python environment (bpy), as this is the only way to access the Blender API.

`blenderproc run <your_python_script>`
```bash
blenderproc run <your_python_script>
```

You can check out this notebook to try BlenderProc in Google Colab, demos the basic examples provided [here](https://github.com/DLR-RM/BlenderProc/tree/main/examples/basics).
Here are some images rendered with the basic example:

![colors](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/colors.png)
![normals](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/normals.png)
![depth](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/depth.png)
![colors](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/colors.png).
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
![normals](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/normals.png).
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
![depth](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/synthetic-data-creation-PBR/depth.png).
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved

## Blender Resources

Expand Down
2 changes: 1 addition & 1 deletion chapters/en/unit10/datagen-diffusion-models.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ This means we have many tools under our belt to generate synthetic data!

## Approaches to Synthetic Data Generation

There are generally three cases for needing synthetic data,
There are generally three cases for needing synthetic data:

**Extending an existing dataset:**

Expand Down
14 changes: 7 additions & 7 deletions chapters/en/unit10/point_clouds.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -22,19 +22,19 @@ The 3D Point Data is mainly used in self-driving capabilities, but now other AI

## Generation and Data Representation

We will be using the python library [point-cloud-utils](https://github.com/fwilliams/point-cloud-utils), and [open-3d](https://github.com/isl-org/Open3D), which can be installed by
We will be using the python library [point-cloud-utils](https://github.com/fwilliams/point-cloud-utils), and [open-3d](https://github.com/isl-org/Open3D), which can be installed by:

```bash
pip install point-cloud-utils
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
```

We will be also using the python library open-3d, which can be installed by
We will be also using the python library open-3d, which can be installed by:

```bash
pip install open3d
```

OR a Smaller CPU only version
OR a Smaller CPU only version:

```bash
pip install open3d-cpu
Expand All @@ -53,13 +53,13 @@ The type of file is inferred from its file extension. Some of the extensions sup

- A simple PLY object consists of a collection of elements for representation of the object. It consists of a list of (x,y,z) triplets of a vertex and a list of faces that are actually indices into the list of vertices.
- Vertices and faces are two examples of elements and the majority of the PLY file consists of these two elements.
- New properties can also be created and attached to the elements of an object, but these should be added in such a way that old programs do not break when these new properties are encountered
- New properties can also be created and attached to the elements of an object, but these should be added in such a way that old programs do not break when these new properties are encountered.

** STL (Standard Tessellation Language) **

- This format approximates the surfaces of a solid model with triangles.
- These triangles are also known as facets, where each facet is described by a perpendicular direction and three points representing the vertices of the triangle.
- However, these files have no description of Color and Texture
- However, these files have no description of Color and Texture.

** OFF (Object File Format) **

Expand All @@ -77,11 +77,11 @@ The type of file is inferred from its file extension. Some of the extensions sup

- X3D is an XML based 3D graphics file format for presentation of 3D information. It is a modular standard and is defined through several ISO specifications.
- The format supports vector and raster graphics, transparency, lighting effects, and animation settings including rotations, fades, and swings.
- X3D has the advantage of encoding color information (unlike STL) that is used during printing the model on a color 3D printer
- X3D has the advantage of encoding color information (unlike STL) that is used during printing the model on a color 3D printer.

** DAE (Digital Asset Exchange) **

- This is an XML schema which is an open standard XML schema, from which DAE files are built.
- This file format is based on the COLLADA (COLLAborative Design Activity) XML schema which is an open standard XML schema for the exchange of digital assets among graphics software applications
- This file format is based on the COLLADA (COLLAborative Design Activity) XML schema which is an open standard XML schema for the exchange of digital assets among graphics software applications.
- The format's biggest selling point is its compatibility across multiple platforms.
- COLLADA files aren't restricted to one program or manufacturer. Instead, they offer a standard way to store 3D assets.
4 changes: 2 additions & 2 deletions chapters/en/unit10/synthetic-lung-images.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ The generator has the following model architecture:
- Conv2D layer
- Batch Normalization layer
- ReLU activation
- Conv2D layer with Tanh activation
- Conv2D layer with Tanh activation.

The discriminator has the following model architecture:

Expand All @@ -27,7 +27,7 @@ The discriminator has the following model architecture:
- Conv2D layer
- Batch Normalization layer
- Leaky ReLU activation
- Conv2D layer with Sigmoid
- Conv2D layer with Sigmoid.

**Data Collection**

Expand Down
8 changes: 4 additions & 4 deletions chapters/en/unit10/synthetic_datasets.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -39,8 +39,8 @@ Semantic segmentation is vital for autonomous vehicles to interpret and navigate

| Name | Year | Description | Paper | | Additional Links |
|---------------------|--------------|-------------|----------------|---------------------|---------------------|
| Virtual KITTI 2 | 2020 | Virtual Worlds as Proxy for Multi-Object Tracking Analysis | [Virtual KITTI 2](https://arxiv.org/pdf/2001.10773.pdf) | | [Website](https://europe.naverlabs.com/Research/Computer-Vision/Proxy-Virtual-Worlds/) |
| ApolloScape | 2019 | Compared with existing public datasets from real scenes, e.g. KITTI [2] or Cityscapes [3], ApolloScape contains much large and richer labeling including holistic semantic dense point cloud for each site, stereo, per-pixel semantic labeling, lane-mark labeling, instance segmentation, 3D car instance, high accurate location for every frame in various driving videos from multiple sites, cities, and daytimes | [The ApolloScape Open Dataset for Autonomous Driving and its Application](https://arxiv.org/abs/1803.06184) | | [Website](https://apolloscape.auto/) |
| Virtual KITTI 2 | 2020 | Virtual Worlds as Proxy for Multi-Object Tracking Analysis. | [Virtual KITTI 2](https://arxiv.org/pdf/2001.10773.pdf) | | [Website](https://europe.naverlabs.com/Research/Computer-Vision/Proxy-Virtual-Worlds/) |
sergiopaniego marked this conversation as resolved.
Show resolved Hide resolved
| ApolloScape | 2019 | Compared with existing public datasets from real scenes, e.g. KITTI [2] or Cityscapes [3], ApolloScape contains much large and richer labeling including holistic semantic dense point cloud for each site, stereo, per-pixel semantic labeling, lane-mark labeling, instance segmentation, 3D car instance, high accurate location for every frame in various driving videos from multiple sites, cities, and daytimes. | [The ApolloScape Open Dataset for Autonomous Driving and its Application](https://arxiv.org/abs/1803.06184) | | [Website](https://apolloscape.auto/) |
| Driving in the Matrix | 2017 | The core idea behind "Driving in the Matrix" is to use photo-realistic computer-generated images from a simulation engine to produce annotated data quickly. | [Driving in the Matrix: Can Virtual Worlds Replace Human-Generated Annotations for Real World Tasks?](https://arxiv.org/pdf/1610.01983.pdf) | | [GitHub](https://github.com/umautobots/driving-in-the-matrix) ![GitHub stars](https://img.shields.io/github/stars/umautobots/driving-in-the-matrix.svg?style=social&label=Star) |
| CARLA | 2017 | **CARLA** (CAR Learning to Act) is an open simulator for urban driving, developed as an open-source layer over Unreal Engine 4. Technically, it operates similarly to, as an open source layer over Unreal Engine 4 that provides sensors in the form of RGB cameras (with customizable positions), ground truth depth maps, ground truth semantic segmentation maps with 12 semantic classes designed for driving (road, lane marking, traffic sign, sidewalk and so on), bounding boxes for dynamic objects in the environment, and measurements of the agent itself (vehicle location and orientation). | [CARLA: An Open Urban Driving Simulator](https://arxiv.org/pdf/1711.03938v1.pdf) | | [Website](https://carla.org/) |
| Synthia | 2016 | A large collection of synthetic images for semantic segmentation of urban scenes. SYNTHIA consists of a collection of photo-realistic frames rendered from a virtual city and comes with precise pixel-level semantic annotations for 13 classes: misc, sky, building, road, sidewalk, fence, vegetation, pole, car, sign, pedestrian, cyclist, lane-marking. | [The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes](https://www.cv-foundation.org/openaccess/content_cvpr_2016/html/Ros_The_SYNTHIA_Dataset_CVPR_2016_paper.html) | | [Website](https://synthia-dataset.net/) |
Expand All @@ -55,8 +55,8 @@ Navigating indoor environments can be challenging due to their complexity. These
| Name | Year | Description | Paper | Additional Links |
|--------------|--------------|-------------|----------------|--------------|
|Habitat | 2023 | An Embodied AI simulation platform for studying collaborative human-robot interaction tasks in home environments. | [HABITAT 3.0: A CO-HABITAT FOR HUMANS, AVATARS AND ROBOTS](https://ai.meta.com/static-resource/habitat3) | [Website](https://aihabitat.org/habitat3/) |
| Minos | 2017 | Multimodal Indoor Simulator | [MINOS: Multimodal Indoor Simulator for Navigation in Complex Environments](https://arxiv.org/pdf/1712.03931.pdf) | [GitHub](https://github.com/minosworld/minos) ![GitHub stars](https://img.shields.io/github/stars/minosworld/minos.svg?style=social&label=Star) |
| House3D | 2017 (archived in 2021) | A Rich and Realistic 3D Environment | [Building generalisable agents with a realistic and rich 3D environment](https://arxiv.org/pdf/1801.02209v2.pdf) | [GitHub](https://github.com/facebookresearch/House3D) ![GitHub stars](https://img.shields.io/github/stars/facebookresearch/House3D.svg?style=social&label=Star) |
| Minos | 2017 | Multimodal Indoor Simulator. | [MINOS: Multimodal Indoor Simulator for Navigation in Complex Environments](https://arxiv.org/pdf/1712.03931.pdf) | [GitHub](https://github.com/minosworld/minos) ![GitHub stars](https://img.shields.io/github/stars/minosworld/minos.svg?style=social&label=Star) |
| House3D | 2017 (archived in 2021) | A Rich and Realistic 3D Environment. | [Building generalisable agents with a realistic and rich 3D environment](https://arxiv.org/pdf/1801.02209v2.pdf) | [GitHub](https://github.com/facebookresearch/House3D) ![GitHub stars](https://img.shields.io/github/stars/facebookresearch/House3D.svg?style=social&label=Star) |


### Human Action Recognition and Simulation
Expand Down
2 changes: 1 addition & 1 deletion chapters/en/unit12/conclusion.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ This is work that highlights and explores techniques for making machine learning
### 🧑‍🤝‍🧑 Inclusive

These are projects which broaden the scope of who builds and benefits in the machine learning world. Some examples:
- Curating diverse datasets that increase the representation of underserved groups
- Curating diverse datasets that increase the representation of underserved groups.
- Training language models on languages that aren't yet available on the Hugging Face Hub.
- Creating no-code and low-code frameworks that allow non-technical folk to engage with AI.

Expand Down
6 changes: 3 additions & 3 deletions chapters/en/unit13/hyena.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -91,8 +91,8 @@ Some work has been conducted to speed up this computation like FastFFTConv based

![nd_hyena.png](https://huggingface.co/datasets/hf-vision/course-assets/resolve/main/outlook_hyena_images/nd_hyena.png)
In essence, Hyena can be performed in two steps:
1. Compute a set of N+1 linear projections similarly of attention (it can be more than 3 projections)
2. Mixing up the projections: The matrix \\(H(u)\\) is defined by a combination of matrix multiplications
1. Compute a set of N+1 linear projections similarly of attention (it can be more than 3 projections).
2. Mixing up the projections: The matrix \\(H(u)\\) is defined by a combination of matrix multiplications.

## Why Hyena Matters

Expand All @@ -113,7 +113,7 @@ Hyena has been applied to N-Dimensional data with the Hyena N-D layer and can be
here is a noticeable enhancement in GPU memory efficiency with the increase in the number of image patches.

Hyena Hierarchy facilitates the development of larger, more efficient convolution models for long sequences.
The potential for Hyena type models for computer vision would be a more efficient GPU memory consumption of patches, that would allow :
The potential for Hyena type models for computer vision would be a more efficient GPU memory consumption of patches, that would allow:
- The processing of larger, higher-resolution images
- The use of smaller patches, allowing a fine-graine feature representation

Expand Down
Loading
Loading