Handloomed fabrics recognition with deep learning Scientific Reports

0 views

AI-based histopathology image analysis reveals a distinct subset of endometrial cancers Nature Communications

ai based image recognition

This can result in significant cost savings and faster time-to-market for new products and features. In e-commerce, transfer learning can be used to improve product search and recommendation systems, automate product tagging and categorization, and enable visual search capabilities. Transfer ai based image recognition learning can also be used to improve image and video analysis for tasks such as product quality control and visual inspection. From the literature, most authors use a few thousand images for training models, and it highlights the need for more data for specific vegetable diseases.

Consequently, we integrated this method into our comparison framework, referring to it as HED for clarity and consistency. Additionally, we employed the Macenko method as a standalone color normalization approach using only one reference image. For the lithology segmentation and recognition part of this study, we accurately annotate rock lithology images based on source information, covering rock attributes such as porphyrite, granite, loess clay, fault, and background.

ai based image recognition

The CNN model outperforms all other models in accuracy tests, reaching an impressive 99.62% (Table 9). In this paper (Kanaparthi and Ilango, 2023), DL methods investigated the training issues on the Chilli leaf diseases dataset. This research uses 160 images from the public domain repository on Kaggle to assess the efficacy of the Squeeze-Net training architecture in identifying Geminivirus and Mosaic-infected Chilli leaves. Training accuracy varies from 50% to 100% as a function of settings like CNN optimizers, Max-epochs, dropout probability, strides, dilation factor, and padding values. Adopting Adam and RMSprop optimizers with epochs of 40 and 35, respectively, leads to a perfect accuracy score for the Squeeze-Net CNN architecture (Lin et al., 2019a) and achieves 100% accuracy.

One important aspect of chest X-ray positioning is the area of the X-ray field relative to the patient’s chest34,35. During acquisition, this area may be ‘collimated’ in order to cover the relevant anatomy while limiting unnecessary X-ray exposure to other regions34,35,36. After acquisition, the image may also be ‘electronically collimated’ via cropping37,38. These adjustments effectively alter the field of view of the image, and this parameter is the second factor we consider.

What is AI? Everything to know about artificial intelligence

The curve takes shape around this point, illustrating the performance of the model across different thresholds26. Edenphotos is an AI-powered image storage and organization solution that provides users with an intuitive and efficient way to manage their digital photos. The platform automatically tags photos using advanced image recognition technology and categorizes them into relevant themes and situations. This ensures that users can easily find and access their photos, without the need for manual sorting.

Pre-trained deep learning models for brain MRI image classification – Frontiers

Pre-trained deep learning models for brain MRI image classification.

Posted: Wed, 26 Jun 2024 02:55:15 GMT [source]

The experimental results showed that Residual network-50 performed more reliably in terms of accuracy, sensitivity and specificity values14. Jacob and Darney designed a CNN-based IR model to improve the accuracy of IR in IoT, ChatGPT App and evaluated the experiments on the IoT image dataset practical appropriateness in IoT systems15. Investigation of histopathology slides by pathologists is an indispensable component of the routine diagnosis of cancer.

Zhang et al. (2018) designed the RefineDet algorithm, which inherited the advantages of single-stage detectors and two-stage detectors. RefineDet uses VGG-16 or ResNet-101 as the backbone network for feature extraction, and integrates the neck structure (feature pyramid and feature fusion) into the head structure. Goodfellow et al. (2014) proposed Generative Adversarial Networks (GANs), which are unsupervised generative models that work based on the maximum ChatGPT likelihood principle and use adversarial training. The objective behind adversarial learning is to train the detection network by using an adversarial network to generate occlusion and deformed image samples, and it is one of the most used generative model methods for generating data distribution. GAN is more than just an image generator; it also uses training data to perform object detection, segmentation, and classification tasks across various domains.

Deploying and scaling distributed parallel deep neural networks on the Tianhe-3 prototype system

It is important as the scenario of false negative in this case, i.e. predicting a powerloom “gamucha” as handloom has significant effect. Similarly, a high recall ensures that the model does not miss important instances of the positive class. Our lightweight model demonstrates remarkable performance while maintaining computational efficiency, marking a significant achievement, especially considering its intended integration into a smartphone application. These images were identified and checked and were found to be blurry, indicating that the images have to be well focused before running the model. It is used to develop cross-platform applications for Android, iOS, Linux, macOS, Windows, Google Fuchsia, and the web from a single codebase. The production of high-quality handloom “gamucha” demands significant skill and time from weavers, resulting in a meticulous process.

ai based image recognition

Computer vision involves a wide range of techniques and approaches, enabling models to learn from large amounts of visual data, such as images and videos. There have been many recent achievements in computer vision, driven in large part by advances in deep learning and neural networks. Our work adds to the growing attention towards better understanding the underlying causes of AI bias and behavior across protected subgroups1,2,7,8,42,45,52. In the current context, it has been suggested that factors ranging from demographic confounders to label bias42,43,44 could contribute to the performance differences observed by Seyyed-Kalantari et al.1.

The first set of models are trained to predict self-reported race based on chest X-ray images (Fig. 1a). We then examine how the predictions of these models change when varying several technical parameters. We use the resulting knowledge to inform the development of a second set of models.

As discussed above, various vegetable diseases have limited data and non-uniformity between the classes. To prevent bias, it’s vital to represent diseases by vegetable samples of similar size, both infected and healthy, to maintain a balanced and unbiased dataset for accurate analysis and prediction. In this study (Arshaghi et al, 2023), machine vision and AI identify defects in agricultural goods like potatoes. Potato diseases include healthy, black scurf, common scab, black leg, and pink rot. Compared to previous approaches, the accuracy of the suggested DL methodology was much more significant, reaching 100% and 99% in various disease groups (Table 9).

Quantification and statistical analysis

Version 2.0 will also include View Finder Gamma Display Assist while using S-Log3 for monitoring. BURANO Version 2.0 will also add 1.8x de-squeeze setting as well as additional high frame rate (S & Q) modes including 66, 72, 75, 88, 90, 96, 110 fps. Planned to be released in March 2025, BURANO Version 2.0 offers many new features and improvements requested by the user community, including new recording formats, new 1.8x de-squeeze, and monitoring improvements. You can foun additiona information about ai customer service and artificial intelligence and NLP. So, to help you plan accordingly, here is Sony’s full roadmap for its Cinema LIne including what feature upgrades are set to come and when these new firmware updates should be released. These reports and examples might be fascinating from a technical perspective, but we’ll still need to see this new AI algorithm implemented into cameras before we can say how much it will revolutionize the industry.

Throughout the text, ‘95% CI’ was used when representing the 95% confidence interval and ‘±’ was used when representing standard deviation. A.Z., A.C., M.K., D.F., D.G.H., A.C., P.B., G.W., and C.B.G. contributed pathology expertise. All the authors critically reviewed the manuscript for important intellectual content and approved the final manuscript. 12, the on-site engineering team conducted laboratory tests on rock samples collected from the field. The laboratory test results were compared with the RC values predicted using the correction factor method. The results show that the Transformer + UNet model’s success rate is as high as 95.57%, surpassing other popular models such as DeepLabV3, DeepLabV3 + , FPN, Linknet, PSPNet, PAN, and UNet +  + .

A novel boosted ada-boost classifier for MRI-based brain tumour detection

To overcome this challenge, adversarial domain adaptation networks have been employed, however, these networks tend to decrease the discriminability of the learned features and do not fully utilize the knowledge transferability of the target domain. To address these shortcomings, we proposed an approach referred to as AIDA, which enhances the adversarial domain adaptation network using the frequency domain information through an FFT-Enhancer module. By integrating the color space of target domain samples into the label prediction loss, our approach effectively addressed the challenge of overfitting the network to the source domain. This integration yielded significant benefits, as the network demonstrated enhanced generalization capabilities, enabling it to more accurately classify the target domain. Consequently, our approach surpassed the limitations of previous methods by improving the network’s discriminability for both the source and target domains.

The model’s mAP is 1.9 percentage points higher than that of the original RetinaNet, indicating improved detection accuracy. Additionally, in scenarios where electrical equipment is densely arranged at various angles, the rotating rectangular frame achieves more precise detection than the horizontal frame, as illustrated in Fig. The larger the AG, the richer the information of edge texture is represented, and the comparison of AG of each algorithm is shown in Table 1. The Ani-SSR, by preserving more image details while enhancing contrast, exhibits an improvement in the average gradient score compared to the other three algorithms, objectively demonstrating the effectiveness of the proposed algorithm in this paper. The acquisition of temperature information for substation electrical equipment largely depends on infrared thermography (IRT).

Development of OrgaExtractor as a deep learning-based organoid image processing tool

Early experiments with the new AI have shown that the recognition accuracy exceeds conventional methods and is powered by an algorithm that can classify objects based on their appearances. In the report, Panasonic lists examples of these categories as “train” or “dog” as well as subcategories as “train type” or “dog breed” based on different appearances. The Cap is prone to current-heating faults, often due to internal bolt loosening or wiring aging corrosion and other reasons that increase the resistance, resulting in an increase in the amount of heat generated. Initial detection of Cap is carried out using improved RetinaNet, and the results are input into DeeplabV3 + model for segmentation, thus separating n regions of the Cap. The local temperature maximum T1, T2, T3…Tn are yielded, the maximum value is selected as the hot spot temperature Tmax and the minimum value is selected as the normal temperature Tmin, and the relative temperature difference δt is obtained. If the Tmax and δt satisfy the discriminating conditions, it is determined as the corresponding fault level, and if they do not satisfy the conditions, it is judged that the equipment is normal.

  • Additionally, the Path Aggregation Network (PAN) module and an Attention module have been incorporated into the feature fusion stage of the original RetinaNet.
  • Essentially, we’re talking about a system or machine capable of common sense, which is currently unachievable with any available AI.
  • This research aims to introduce a unique Global Pooling Dilated CNN (GPDCNN) for plant disease identification (Zhang et al., 2019).
  • As the baseline architecture for our classifier, we exploited ResNet1844, a simple and effective residual network, with the pre-trained ImageNet45 weights.

Here, the study aimed to identify defects in a handloom silk fabric using image analysis techniques. The disparity in sensitivity of the AI diagnostic model was quantified as the sensitivity of the model for white patients minus the sensitivity of the model for patients of other races. Error bars correspond to standard deviation computed via bootstrapping and are plotted with respect to the point estimate in the MXR test split. The results are derived from 1992, 10,335, and 38,282 images for Asian, Black, and white patients respectively. The other technical factors we explore relate to the positioning of the patient.

  • Specifically, all layers’ connection architecture is employed, i.e., each layer acquires inputs from all previous layers and conveys its own feature maps to all subsequent layers.
  • While subtle, this effectively changes the overall contrast within the image, such as the relative difference in intensity between lung and bone regions.
  • All other confidence intervals, standard deviations, and p-values were computed via bootstrapping with 2000 samples.

2 then represent the percent change in average prediction scores per race for each preprocessing combination compared to the original processing. The average scores of the racial identity prediction model were computed for different window width and field of view values and compared to the default preprocessing used to train the model. The average scores were computed in a weighted fashion to equally weight each patient race across the test dataset (see “Methods”). A positive change (red) indicates an increase in the average score for the corresponding race and preprocessing combination across the entire test set.

The complexity of classroom discourse can be measured by the length of sentences spoken. Based on Table 2, this work selects Mandarin clarity as an evaluation indicator for CDA in online courses, serving as a fundamental feature of classroom discourse. Test results of models trained on PTB-XL ECGs and tested on a holdout test set from PTB-XL. About LEAFIO AIThe LEAFIO AI Retail Automation Platform empowers retailers with robust, agile, and adaptable automation technologies.

From:
Date: October 15, 2024

Store Product Category