Using toy simulations to understand the Statistical Learning Theory " Framework " for Supervised Classification. youtube.com/watch?v=lsYPC…
πŸ‘︎ 7
πŸ’¬︎
πŸ‘€︎ u/niszoig
πŸ“…︎ Jan 07 2022
🚨︎ report
I am planning to do a meta-analysis tutorial after I finish my current text mining and supervised text classification tutorials. Anyone interested in meta-analysis in R? All tutorials will be in YouTube. Channel: R_Py Data Science. Example video link: youtu.be/4I3PCDME5U8
πŸ‘︎ 23
πŸ’¬︎
πŸ‘€︎ u/aabush1
πŸ“…︎ Dec 12 2021
🚨︎ report
Performance or Trust? Why Not Both. Deep AUC Maximization with Self-Supervised Learning for COVID-19 Chest X-ray Classifications by Siyuan He et al. deepai.org/publication/pe…
πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/deep_ai
πŸ“…︎ Dec 19 2021
🚨︎ report
Improving Supervised Classification of Suburban Parks

Hey everyone!

I'm trying to quantify how much the Urban Tree Canopy (UTC) has changed within parks for a local municipality. Using ArcGIS Pro and orthoimagery, I've run through a supervised classification using the Imagery Wizard and using the Classification tool. In both cases it seems that my Water category and my Forest classes are being confused, as well as some slight mixing of the other classes.

I was curious if anyone had some suggestions for how to minimize this? I've created several (14) classes within my schema and tried to create 20-30 training samples for any classes that were included in the imagery. Open to any suggestions for how to produce less "muddled" land use classification. Thanks!

πŸ‘︎ 40
πŸ’¬︎
πŸ‘€︎ u/mmoy8684
πŸ“…︎ Oct 12 2021
🚨︎ report
(Supervised Learning; Multiclass Classification) If intercoder reliability is an issue, should it lower our expectations for model performance?

Suppose there's a set of (multiclass) labeled documents, say 10 classes, and the team found intercoder reliability to be about 90%. I could account for this directly in PyTorch's nn.CrossEntropyLoss by defining label_smooth=0.1 to give the model soft rather than hard targets.

Beyond this though, what kind of performance should I interpret as "good enough"? Right now, with that stated intercoder reliability, my model is at approximately 89.8% accuracy / 90% F1 score. There are some examples of duplicate text entries in the training data that are coded differently, i.e. conditioned on two documents having the same input text, they get coded as different labels roughly 5.3% of the time.

Should I just expect that 90% is about as good as it gets, since the labels themselves slightly weaker than would be the case in "pure" supervised learning?

πŸ‘︎ 7
πŸ’¬︎
πŸ‘€︎ u/eadala
πŸ“…︎ Nov 13 2021
🚨︎ report
For topic classification, what's the difference between supervised & semi-supervised learning?

In particular, say I'm using some transformer architecture like BERT, and am fine-tuning it for a downstream task like multiclass classification of documents into 1 of 5 topics. I have 10,000 labeled instances in my data, and 40,000 unlabeled instances. If I wanted to handle this in a supervised fashion, I would maybe:

  • Train the model on 8,000 of the labeled instances, and validate on 1,000 of them, with the remaining 1,000 set aside as test set.
  • Tinker with the model until performance on the validation set is as good as I can get it.
  • Cross-validate by re-constructing the 8,000-1,000 split (never touching the 1,000 test set).
  • Repeat the tinkering process until cross-validated average performance is as good as I can get it.
  • Test on the 1,000 test set.
  • If the performance is adequate, I "trust" the model and predict the topics for the remaining 40,000 instances.

I am a bit confused how semi-supervised learning would differ in this sense. As I understand it, I'm still going through a training process. Would it be something like this?

  • Train on 8,000, validate on 1,000. Cross-validate & repeat tinkering as above.
  • Predict remaining 40,000 instances.
  • Retrain the model on these 49,000 instances split in some way for training & validation.
  • Assign extra penalty (greater loss) for mis-classifying the instances for which ground-truth labels exist.

Sorry if that's completely wrong; just thinking out loud. Is semi-supervised learning something that would be appropriate in this setting? What would it look like?

πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/eadala
πŸ“…︎ Oct 25 2021
🚨︎ report
I have to carry out supervised classifications of satellite images of Cuprite, Nevada on multispec but I am confused about how I get the images to look bright and colourful so I can make different fields for the alteration minerals. I'm a beginner and just very confused. reddit.com/gallery/pl4x50
πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/katiekunt
πŸ“…︎ Sep 09 2021
🚨︎ report
Study demonstrates the quantum speed up of supervised machine learning on a new classification task phys.org/news/2021-08-qua…
πŸ‘︎ 10
πŸ’¬︎
πŸ‘€︎ u/izumi3682
πŸ“…︎ Aug 26 2021
🚨︎ report
Image classification using self-supervised learning

I studied and implemented modern ConvNets for image classification task. Now I want to classify images using slef-supervised learning methods. I don't know about self-supervised learning. Where should I start from? And what track track should I follow to master self-supervised learning methods? Please recommend me books, papers, and blogs so that I can follow it.

πŸ‘︎ 20
πŸ’¬︎
πŸ‘€︎ u/SAbdusSamad
πŸ“…︎ Jun 21 2021
🚨︎ report
Signature Editor for Supervised Classification

Everytime i try to use the Signature Editor in Erdas Imagine V. 16.6.0 Build 2100, it stops working. Anyone else having this issue?

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/P_S_P_S
πŸ“…︎ Aug 12 2021
🚨︎ report
How can supervised classification algorithms be used in business?

I'm writing a paper on machine learning and I'd like to list a few examples of business applications for classification. So far I have sentiment analysis, document classification and recommendations. Are there any more that would be good to include or is that pretty much it?

Thanks in advance.

πŸ‘︎ 6
πŸ’¬︎
πŸ‘€︎ u/lifelifebalance
πŸ“…︎ May 05 2021
🚨︎ report
[R] Semi-Supervised Speech Recognition via Graph-based Temporal Classification arxiv.org/abs/2010.15653
πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/fasttosmile
πŸ“…︎ Jun 29 2021
🚨︎ report
[R] Big Self-Supervised Models Advance Medical Image Classification

A new work from Google Brain (authors of SimCLR) and Google Health shows self-supervised pretraining on unlabeled medical images is much more effective than supervised pretraining on ImageNet.

https://arxiv.org/pdf/2101.05224v1.pdf

πŸ‘︎ 125
πŸ’¬︎
πŸ‘€︎ u/aifordummies
πŸ“…︎ Jan 15 2021
🚨︎ report
Can someone explain to be training process for semi-supervised classification task

Hi,

I am going through the Graph convolutional neural network paper for semisupervised label classification (https://arxiv.org/pdf/1609.02907.pdf ).

The github repository for the same is here [https://github.com/tkipf/pygcn/blob/master/pygcn/train.py]

What I do not understand is how they do semi-supervised training. I see that they input all the data but train with only few examples. They at some-point mask the labels. This is the part that I am not getting.

Should I mask all labels other than the training examples? Then what am I testing against? Can someone who has done this before please clarify?

πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/popkept09
πŸ“…︎ May 25 2021
🚨︎ report
Autoencoder supervised classification

I just want to know from the professionals that how to perform supervised classification by auto encoder My data has 500 sensor variables and a class label which defines abnormal and normal class of the system. If someone gives me codes to implement autoencoder in Keras will be highly appreciable

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/Few_Echidna8400
πŸ“…︎ May 15 2021
🚨︎ report
Supervised classification classes help

Hi,

I am currently trying to determine which classes to use in my supervised classifications. I have a pre and post sentinel-2 image https://imgur.com/a/MP5utAT of an area of Victoria affected by the 2019-2020 wildfires. I have completed my pre-processing and changed the spectral bands to R: Band 12, G: Band 8 and B: Band 4 which provides good visibility to vegetation classes, highlighting the burnt areas. I currently have a list of classes which include;

  • Water
  • Sand
  • Cloud
  • Impervious surfaces
  • Tree
  • Scrub
  • Bare soil
  • Burn scar
  • Grassland
  • Vegetation

I currently have a few issues with these classes as scrub (Dark purple), bare soil (Light pink) and burn scar (Purple/Pink) all seem to have a similar spectral reflectance and it could make distinguishing between them difficult when creating my training classes and for the computer when creating the classification. I wondering if there's any spectral band combinations that will make it easier to differentiate between them? I also have the same issue with sand, cloud and impervious surfaces which all have a spectral reflectance of White.

I'm also wondering if I've missed any obvious classes off to include?

Thanks

πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/BigPurpleAki
πŸ“…︎ Jan 15 2021
🚨︎ report
"Big Self-Supervised Models Advance Medical Image Classification", Azizi et al 2021 (SimCLR pretraining on medical datasets) /r/MachineLearning/commen…
πŸ‘︎ 7
πŸ’¬︎
πŸ‘€︎ u/gwern
πŸ“…︎ Jan 15 2021
🚨︎ report
Hello everyone! I'm new to RS. I'd like to know if it's possible to create 3D maps from the output images of the supervised classification process. If so, how can I do this? Thanks in advance!
πŸ‘︎ 9
πŸ’¬︎
πŸ‘€︎ u/dbensons
πŸ“…︎ Jul 18 2020
🚨︎ report
Month Five of the Bot Classification Project. It's a much lengthier video including my thoughts throughout each step in the end-to-end machine learning process from data collection through supervised learning. Questions/comments always appreciated! (context in comments for the unfamiliar) youtu.be/hzCpwPTXzrA
πŸ‘︎ 25
πŸ’¬︎
πŸ‘€︎ u/chriskok1337
πŸ“…︎ Jul 06 2020
🚨︎ report
SECANT: a biology-guided semi-supervised method for clustering, classification, and annotation of single-cell multi-omics biorxiv.org/content/10.11…
πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/sburgess86
πŸ“…︎ Nov 09 2020
🚨︎ report
[D] Unsupervised Linear Classification in Semi-Supervised learning is misleading!

TLDR: I mean we need a new name, why do we even call it unsupervised if we use the entire labelled dataset.

If I misunderstood anything here, please do correct me.

From what I have understood is that, in unsupervised linear classification we use the feature extractor from semi-supervised learning and place a linear classifier on top of it. We keep the feature extractor's parameters the same, and only allow the linear classifier to update it's params on the entire training set.

The word unsupervised should not be used here, we need a new word, in my opinion.

What do you guys think?

πŸ‘︎ 6
πŸ’¬︎
πŸ‘€︎ u/EhsanSonOfEjaz
πŸ“…︎ Jun 01 2020
🚨︎ report
"Billion-scale semi-supervised learning for state-of-the-art image and video classification", Facebook ai.facebook.com/blog/bill…
πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/gwern
πŸ“…︎ Oct 30 2020
🚨︎ report
[Research] Self-Supervised Prototypical Transfer Learning for Few-Shot Classification

https://arxiv.org/abs/2006.11325

[pdf] [code]

Summary: "ProtoTransfer" method: metric self-supervised pre-training (ProtoCLR) combined with a matching few-shot transfer-learning approach (ProtoTune). On mini-ImageNet, ProtoTransfer outperforms all state-of-the-art un/self-supervised few-shot approaches (by 4% to 8%). Competitive with fully supervised performance (0% to 4% gap) on 4 cross-domain datasets, at a fraction of the label cost (<1%).

Self-Supervised Prototypical Transfer Learning (ProtoTransfer). 2 stages: (a) self-supervised prototypical pretraining (ProtoCLR) and (b) prototypical fine-tuning & inference (Prototune)

Generalization gap observation: Negligible generalization gap from training classes to test classes (from the same class distribution, e.g. mini-ImageNet). Other supervised & self-supervised few-shot approaches, such as ProtoNet (Snell et al., 2017) and UMTRA (Khodadadeh et al., 2019), respectively, show non-negligible generalization gaps.

e.g. 5-way 5-shot:

Method Train accuracy (%) Test accuracy (%) Generalization gap (%)
ProtoNet 79.09 Β± 0.69 66.33 Β± 0.68 12.76
UMTRA(-ProtoNet) 56.43 Β± 0.78 53.37 Β± 0.68 3.06
ProtoCLR-ProtoNet (this paper) 63.47 Β± 0.58 63.35 Β± 0.54 0.12
πŸ‘︎ 9
πŸ’¬︎
πŸ‘€︎ u/ArnoutDevos
πŸ“…︎ Jul 29 2020
🚨︎ report
Help: Image classification using self-supervised learning

I studied and implemented modern ConvNets for image classification task. Now I want to classify images using slef-supervised learning methods. I don't know about self-supervised learning. Where should I start from? And what track track should I follow to master self-supervised learning methods? Please recommend me books or papers or blogs.

πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/SAbdusSamad
πŸ“…︎ Jun 21 2021
🚨︎ report
How to do training for semi-supervised learning for classification task?

Hi,

I am struggling to understand the training part of this paper by Thomas Kipf [https://arxiv.org/pdf/1609.02907.pdf ]. The github repo is here [ https://github.com/tkipf/pygcn/blob/master/pygcn/train.py ].

What I do not understand what is happening with masking.

I input the whole data, but use a small portion of labeled data to train. Here should I mask the rest of the data?

What will be my test set then?

Can someone who has worked on this before please guide me through?

πŸ‘︎ 4
πŸ’¬︎
πŸ‘€︎ u/popkept09
πŸ“…︎ May 25 2021
🚨︎ report
Big Self-Supervised Models Advance Medical Image Classification

A new work from Google Brain (authors of SimCLR) and Google Health shows self-supervised pretraining on unlabeled medical images is much more effective than supervised pretraining on ImageNet.

They also propose a new method called Multi-Instance Contrastive Learning (MICLe), which uses multiple images of the same underlying pathology per patient case, when available, to construct more informative positive pairs for self-supervised learning.

https://arxiv.org/pdf/2101.05224v1.pdf

πŸ‘︎ 11
πŸ’¬︎
πŸ‘€︎ u/aifordummies
πŸ“…︎ Jan 15 2021
🚨︎ report
Big Self-Supervised Models Advance Medical Image Classification

A new work from Google Brain (authors of SimCLR) and Google Health shows self-supervised pretraining on unlabeled medical images is much more effective than supervised pretraining on ImageNet.

They also propose a new method called Multi-Instance Contrastive Learning (MICLe), which uses multiple images of the same underlying pathology per patient case, when available, to construct more informative positive pairs for self-supervised learning.

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/aifordummies
πŸ“…︎ Jan 15 2021
🚨︎ report

Please note that this site uses cookies to personalise content and adverts, to provide social media features, and to analyse web traffic. Click here for more information.