Skip to content

Commit 5717606

Browse files
authored
Merge pull request #567 from microsoft/staging
Sync master <-> staging
2 parents 44c0906 + af63f87 commit 5717606

File tree

96 files changed

+6863
-12185
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

96 files changed

+6863
-12185
lines changed

README.md

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,8 @@
11
<img src="scenarios/media/logo_cvbp.png" align="right" alt="" width="300"/>
22

33
```diff
4-
+ March 27: Released v1.1 with new and improved
5-
+ functionality for image retrieval, object detection,
6-
+ keypoint detection and action recognition.
7-
+ For additional details, please refer to our releases page.
4+
+ Update June 24: Added action recognition as new core scenario.
5+
+ Object tracking coming soon (in 2-4 weeks).
86
```
97

108
# Computer Vision
@@ -56,10 +54,10 @@ The following is a summary of commonly used Computer Vision scenarios that are c
5654
| [Detection](scenarios/detection) | Base | Object Detection is a technique that allows you to detect the bounding box of an object within an image. |
5755
| [Keypoints](scenarios/keypoints) | Base | Keypoint detection can be used to detect specific points on an object. A pre-trained model is provided to detect body joints for human pose estimation. |
5856
| [Segmentation](scenarios/segmentation) | Base | Image Segmentation assigns a category to each pixel in an image. |
59-
| [Action recognition](contrib/action_recognition) | Contrib | Action recognition to identify in video/webcam footage what actions are performed (e.g. "running", "opening a bottle") and at what respective start/end times.|
57+
| [Action recognition](scenarios/action_recognition) | Base | Action recognition to identify in video/webcam footage what actions are performed (e.g. "running", "opening a bottle") and at what respective start/end times. We also implemented the i3d implementation of action recognition that can be found under (contrib)[contrib]. |
6058
| [Crowd counting](contrib/crowd_counting) | Contrib | Counting the number of people in low-crowd-density (e.g. less than 10 people) and high-crowd-density (e.g. thousands of people) scenarios.|
6159

62-
We separate the supported CV scenarios into two locations: (i) **base**: code and notebooks within the "utils_cs" and "scenarios" folders which follow strict coding guidelines, are well tested and maintained; (ii) **contrib**: code and other assets within the "contrib" folder, mainly covering less common CV scenarios using bleeding edge state-of-the-art approaches. Code in "contrib" is not regularly tested or maintained.
60+
We separate the supported CV scenarios into two locations: (i) **base**: code and notebooks within the "utils_cv" and "scenarios" folders which follow strict coding guidelines, are well tested and maintained; (ii) **contrib**: code and other assets within the "contrib" folder, mainly covering less common CV scenarios using bleeding edge state-of-the-art approaches. Code in "contrib" is not regularly tested or maintained.
6361

6462
## Computer Vision on Azure
6563

contrib/README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,11 +8,12 @@ Each project should live in its own subdirectory ```/contrib/<project>``` and co
88
## Scenarios
99
| Directory | Project description | Build status (optional) |
1010
|---|---|---|
11-
| [Action recognition](action_recognition) | Action recognition to identify in video/webcam footage what actions are performed (e.g. "running", "opening a bottle") and at what respective start/end times.| |
1211
| [Crowd counting](crowd_counting) | Counting the number of people in low-crowd-density (e.g. less than 10 people) and high-crowd-density (e.g. thousands of people) scenarios. | [![Build Status](https://dev.azure.com/team-sharat/crowd-counting/_apis/build/status/lixzhang.cnt?branchName=lixzhang%2Fsubmodule-rev3)](https://dev.azure.com/team-sharat/crowd-counting/_build/latest?definitionId=49&branchName=lixzhang%2Fsubmodule-rev3)|
12+
| [Action Recognition with I3D](action_recognition) | Action recognition to identify video/webcam footage from what actions are performed (e.g. "running", "opening a bottle") and at what respective start/end times. Please note, that we also have a R(2+1)D implementation of action recognition that you can find under [scenarios](../sceanrios).| |
1313

1414
## Tools
1515
| Directory | Project description | Build status (optional) |
1616
|---|---|---|
17+
| [HTML Demo](html_demo) | These files provide an HTML web page that allows users to visualize the output of a deployed computer vision DNN model. Users can improve on and gain insights from their deployed model by uploading query/test images and examining the model results for correctness through the user interface. The interface includes sample query/test images for testing your own model and example output for 3 types of models: image classification, object detection, and image similarity. | |
1718
| [vm_builder](vm_builder) | This script helps users create a single Ubuntu Data Science Virtual Machine with a GPU with the computer vision recipes repo installed and ready to be used. If you find the script to be out-dated or not working, you can create the VM using the Azure portal or the Azure CLI tool with a few more steps. | |
1819
| [vmss_builder](vmss_builder) | This script helps you setup a cluster of virtual machines with the computer vision recipes repo pre-installed using VMSS. This cluster is designed to be temporal, ie to be spun up and torn down. Users for this cluster will be prescribed a username/password/ip. This setup can be used for hands-on / lab sessions when you need to prepare multiple VM environments for a short period.|
Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,17 +1,12 @@
11
# Action Recognition
22

3-
```diff
4-
+ Feb 2020: We are working on moving code from this folder to scenarios\action_recognition.
5-
+ While this work is ongoing, please visit both locations for implementations and documentation.
6-
```
7-
83
This directory contains resources for building video-based action recognition systems.
94

105
Action recognition (also known as activity recognition) consists of classifying various actions from a sequence of frames:
116

127
![](./media/action_recognition2.gif "Example of action recognition")
138

14-
We implemented two state-of-the-art approaches: (i) [I3D](https://arxiv.org/pdf/1705.07750.pdf) and (ii) [R(2+1)D](https://arxiv.org/abs/1711.11248). This includes example notebooks for e.g. scoring of webcam footage or fine-tuning on the [HMDB-51](http://serre-lab.clps.brown.edu/resource/hmdb-a-large-human-motion-database/) dataset.
9+
We implemented two state-of-the-art approaches: (i) [I3D](https://arxiv.org/pdf/1705.07750.pdf) and (ii) [R(2+1)D](https://arxiv.org/abs/1711.11248). This includes example notebooks for e.g. scoring of webcam footage or fine-tuning on the [HMDB-51](http://serre-lab.clps.brown.edu/resource/hmdb-a-large-human-motion-database/) dataset. The latter can be accessed under [scenarios](../scenarios) at the root level.
1510

1611
We recommend to use the **R(2+1)D** model for its competitive accuracy, fast inference speed, and less dependencies on other packages. For both approaches, using our implementations, we were able to reproduce reported accuracies:
1712

@@ -27,5 +22,5 @@ We recommend to use the **R(2+1)D** model for its competitive accuracy, fast inf
2722

2823
| Directory | Description |
2924
| -------- | ----------- |
30-
| [r2p1d](r2p1d) | Scripts for fine-tuning a pre-trained R(2+1)D model on HMDB-51 dataset
31-
| [i3d](i3d) | Scripts for fine-tuning a pre-trained I3D model on HMDB-51 dataset
25+
| [i3d](i3d) | Scripts for fine-tuning a pre-trained I3D model on HMDB-51
26+
dataset. |

contrib/action_recognition/r2p1d/.gitignore

Lines changed: 0 additions & 114 deletions
This file was deleted.

0 commit comments

Comments
 (0)