Home

Awesome

<div align="center">

[NeurIPS 2024] Hawk: Learning to Understand Open-World Video Anomalies

<div align="center">

This is the official repository for Hawk.

Jiaqi Tang^, Hao Lu^, Ruizheng Wu, Xiaogang Xu, Ke Ma, Cheng Fang,

Bin Guo, Jiangbo Lu, Qifeng Chen and Ying-Cong Chen*

^: Equal contribution. *: Corresponding Author.

GitHub license made-for-VSCode Visits Badge

<img src="figs/icon.png" alt="Have eyes like a HAWK!" width="80"> </div> </div>

πŸ” Motivation - Have eyes like a Hawk!

πŸ“’ Updates

▢️ Getting Started

<!-- 1. [Installation](#installation) 2. [Dataset](#dataset) 3. [Configuration](#configuration) 5. [Testing](#Testing) 4. [Training](#Training) -->

πŸͺ’ Installation

<!-- - *Python >= 3.8.2* - *PyTorch >= 1.8.1* - *Install [Polanalyser](https://github.com/elerac/polanalyser) for processing polarization image* ``` pip install git+https://github.com/elerac/polanalyser ``` - *Install other dependencies by* ``` pip install -r requirements.txt ``` -->

πŸ’Ύ Dataset Preparation

🏰 Pretrained Model

<!-- - Google Drive Link for downloading our [Pretrained Model](https://drive.google.com/file/d/13Cn7tX5bFBxsYZG1Haw5VcqhSxWnNzMW/view?usp=sharing) in K-Ford. -->

πŸ”¨ Configuration

<!-- - The configuration files for [`testing`](FilmRemoval/codes/options/test/test.yml) and [`training`](FilmRemoval/codes/options/train/train.yml). - The Test_K_ford option specifies the number of folds for K-fold cross-validation during testing. The data root option specifies the root directory for the dataset, which is set to Dataset. Other configuration settings include learning rate schemes, loss functions, and logger options. ``` datasets: train: name: Reconstruction mode: LQGT_condition Test_K_ford: K10 # remove from training dataroot: /remote-home/share/jiaqi2/Dataset dataroot_ratio: ./ use_shuffle: true n_workers: 0 batch_size: 1 GT_size: 0 use_flip: true use_rot: true condition: image val: name: Reconstruction mode: LQGT_condition_Val Test_K_ford: K10 # for testing dataroot: /remote-home/share/jiaqi2/Dataset dataroot_ratio: ./ condition: image ``` -->

⏳ Testing

<!-- - Modify `dataroot`, `Test_K_ford` and `pretrain_model_G` in [`testing`](FilmRemoval/codes/options/train/test.yml) configuration, then run ``` python test.py -opt ./codes/options/test/test.yml ``` - The test results will be saved to `./results/testset_name`, including `Restored Image` and `Prior`. -->

πŸ–₯️ Training

<!-- - Modify `dataroot` and `Test_K_ford` in [`training`](FilmRemoval/codes/options/train/train.yml) configuration, then run ``` python train.py -opt ./codes/options/train/train.yml ``` - The logs, models and training states will be saved to `./experiments/name`. You can also use `tensorboard` for monitoring for the `./tb_logger/name`. - Restart Training (To add checkpoint in [`training`](FilmRemoval/codes/options/train/train.yml) configuration) ``` path: root: ./ pretrain_model_G: .../experiments/K1/models/XX.pth strict_load: false resume_state: .../experiments/K1/training_state/XX.state ``` -->

⚑ Performance

<!-- Compared with other baselines, our model achieves state-of-the-art performance: > ⭐ **[Table 1] Quantitative evaluation in image reconstruction with 10-fold cross-validation.** > | Methods | PSNR | SSIM | > |---------|------|------| > | SHIQ| 21.58 | 0.7499 | > | Polar-HR| 22.19 | 0.7176 | > | Uformer| 31.68 | 0.9426 | > | Restormer| 34.32 | 0.9731 | > | Ours| 36.48 | 0.9824 | > ⭐ **[Figure 1] Qualitative Evaluation in image reconstruction.** > ![](fig/image-1.png) > ⭐ **[Figure 2-3] Qualitative Evaluation in Industrial Environment. (QR Reading & Text OCR)** > ![](fig/image-2.png) -->

🌐 Citations

The following is a BibTeX reference:

@inproceedings{atang2024hawk,
  title = {Hawk: Learning to Understand Open-World Video Anomalies},
  author = {Tang, Jiaqi and Lu, Hao and Wu, Ruizheng and Xu, Xiaogang and Ma, Ke and Fang, Cheng and Guo, Bin and Lu, Jiangbo and Chen, Qifeng and Chen, Ying-Cong},
  year = {2024},
  booktitle = {Neural Information Processing Systems (NeurIPS)}
}

πŸ“§ Connecting with Us?

If you have any questions, please feel free to send email to jtang092@connect.hkust-gz.edu.cn.

<!-- ## πŸ“œ **Acknowledgment** This work is supported by the National Natural Science Foundation of China (No. 62206068) and the Natural Science Foundation of Zhejiang Province, China under No. LD24F020002. -->