papers:rg:ndungu2023
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revision | |||
| papers:rg:ndungu2023 [2023/10/23 05:32] – asad | papers:rg:ndungu2023 [2023/10/24 06:27] (current) – asad | ||
|---|---|---|---|
| Line 13: | Line 13: | ||
| - Besides classification, | - Besides classification, | ||
| - Table 2 dos not have a dataset created from Sasmal 23. | - Table 2 dos not have a dataset created from Sasmal 23. | ||
| - | - | + | - Most works depend on deep and shallow CNN (ConvNet, mimicks human visual cortex). The deeper (more layers) the better. |
| + | - Regularization techniques (dropout of weakly connected neurons) used to solve overfitting. | ||
| + | - Bowles 21 uses attention-gate layers to suppress irrelevant info. | ||
| + | - Tang 22 uses multidomain multibranch CNN to take multiple inputs. | ||
| + | - Brand 23 aligned the PC of all galaxies with the axis of the coordinate system via PCA. | ||
| + | - Sadeghi 21 uses Zernike polynomials (SVM) to extract image moments. Ntwaetsile 21 uses Haralick features for texture. | ||
| + | - Tang 19 successfully used FIRST-trained model on NVSS, but not the reverse. | ||
| + | - Wang 21 uses SKnet module for attention. Zhang 22 uses SE Net. | ||
| + | - Tiramisu (Pino 21) works best for source extraction, detection, localization and then classification. | ||
| + | - Main drawback of AI is the lack of explainability. | ||
| + | - Another drawback is that we are reducing 4D data cubes into 2D. | ||
papers/rg/ndungu2023.1698060749.txt.gz · Last modified: by asad
