Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Synthesis optimization and adsorption modeling of biochar for pollutant removal via machine learning
W Zhang, R Chen, J Li, T Huang, B Wu, J Ma, Q Wen… - Biochar, 2023 - Springer
Due to large specific surface area, abundant functional groups and low cost, biochar is
widely used for pollutant removal. The adsorption performance of biochar is related to …
widely used for pollutant removal. The adsorption performance of biochar is related to …
Visual recognition with deep nearest centroids
We devise deep nearest centroids (DNC), a conceptually elegant yet surprisingly effective
network for large-scale visual recognition, by revisiting Nearest Centroids, one of the most …
network for large-scale visual recognition, by revisiting Nearest Centroids, one of the most …
Debiased self-training for semi-supervised learning
Deep neural networks achieve remarkable performances on a wide range of tasks with the
aid of large-scale labeled datasets. Yet these datasets are time-consuming and labor …
aid of large-scale labeled datasets. Yet these datasets are time-consuming and labor …
Scribble-supervised medical image segmentation via dual-branch network and dynamically mixed pseudo labels supervision
Medical image segmentation plays an irreplaceable role in computer-assisted diagnosis,
treatment planning and following-up. Collecting and annotating a large-scale dataset is …
treatment planning and following-up. Collecting and annotating a large-scale dataset is …
Cycle: Learning to self-refine the code generation
Pre-trained code language models have achieved promising performance in code
generation and improved the programming efficiency of human developers. However, their …
generation and improved the programming efficiency of human developers. However, their …
Fine-grained classification with noisy labels
Learning with noisy labels (LNL) aims to ensure model generalization given a label-
corrupted training set. In this work, we investigate a rarely studied scenario of LNL on fine …
corrupted training set. In this work, we investigate a rarely studied scenario of LNL on fine …
Don't stop pretraining? make prompt-based fine-tuning powerful learner
Abstract Language models (LMs) trained on vast quantities of unlabelled data have greatly
advanced the field of natural language processing (NLP). In this study, we re-visit the widely …
advanced the field of natural language processing (NLP). In this study, we re-visit the widely …
Innovative approach for predicting daily reference evapotranspiration using improved shallow and deep learning models in a coastal region: A comparative study
Accurate and reliable estimation of Reference Evapotranspiration (ETo) is crucial for water
resources management, hydrological processes, and agricultural production. The FAO-56 …
resources management, hydrological processes, and agricultural production. The FAO-56 …
Prototype-guided pseudo labeling for semi-supervised text classification
Semi-supervised text classification (SSTC) aims at text classification with few labeled data
and massive unlabeled data. Recent works achieve this task by pseudo-labeling methods …
and massive unlabeled data. Recent works achieve this task by pseudo-labeling methods …
Cosst: Multi-organ segmentation with partially labeled datasets using comprehensive supervisions and self-training
Deep learning models have demonstrated remarkable success in multi-organ segmentation
but typically require large-scale datasets with all organs of interest annotated. However …
but typically require large-scale datasets with all organs of interest annotated. However …