Objective detection

  • 文章类型: Journal Article
    目标:我们的目标是开发一种人工智能(AI)系统,用于使用先进的深度学习来检测计算机断层扫描(CT)图像中的尿路结石,能够实时计算结石参数,例如体积和密度。这对治疗决策至关重要。将系统的性能与急诊室(ER)场景中泌尿科医生的性能进行了比较。
    方法:2022年8月至2023年7月接受结石手术的患者的轴向CT图像包括数据集,分成70%用于培训,10%用于内部验证,20%用于测试。两名泌尿科医生和一名AI专家使用Labelig为地面数据注释了石头。YOLOv4架构用于培训,通过RTX4900图形处理单元(GPU)加速。使用CT图像对100例可疑尿石症患者进行外部验证。
    AI系统在39.433个CT图像上进行了训练,其中9.1%为正。该系统实现了95%的准确度,以1:2的正负样本比率达到峰值。在5736张图像的验证集中(482张阳性),准确率保持在95%。遗漏(2.6%)主要是不规则结石。假阳性(3.4%)通常是由于伪影或钙化。使用来自ER的100张CT图像进行的外部验证显示准确率为94%;错过的病例大多是输尿管膀胱交界处结石,不包括在训练集中。人工智能系统在速度上超过了人类专家,在13秒内分析150张CT图像,而泌尿科医师评估为38.6s,正式阅读为23h。AI系统在0.2s内计算石头体积,与泌尿科医生计算的77秒相比。
    结论:我们的人工智能系统,它使用先进的深度学习,在真实临床环境中协助诊断尿石症的准确率为94%,并具有使用标准消费级GPU进行快速诊断的潜力。
    结果:我们开发了一种新的AI(人工智能)系统,该系统可以在CT(计算机断层扫描)扫描中快速准确地检测肾结石。测试表明,该系统非常有效,急诊科真实病例的准确率为94%。它比传统方法快得多,并提供快速可靠的结果,以帮助医生为患者做出更好的治疗决策。
    OBJECTIVE: Our aim was to develop an artificial intelligence (AI) system for detection of urolithiasis in computed tomography (CT) images using advanced deep learning capable of real-time calculation of stone parameters such as volume and density, which are essential for treatment decisions. The performance of the system was compared to that of urologists in emergency room (ER) scenarios.
    METHODS: Axial CT images for patients who underwent stone surgery between August 2022 and July 2023 comprised the data set, which was divided into 70% for training, 10% for internal validation, and 20% for testing. Two urologists and an AI specialist annotated stones using Labelimg for ground-truth data. The YOLOv4 architecture was used for training, with acceleration via an RTX 4900 graphics processing unit (GPU). External validation was performed using CT images for 100 patients with suspected urolithiasis.
    UNASSIGNED: The AI system was trained on 39 433 CT images, of which 9.1% were positive. The system achieved accuracy of 95%, peaking with a 1:2 positive-to-negative sample ratio. In a validation set of 5736 images (482 positive), accuracy remained at 95%. Misses (2.6%) were mainly irregular stones. False positives (3.4%) were often due to artifacts or calcifications. External validation using 100 CT images from the ER revealed accuracy of 94%; cases that were missed were mostly ureterovesical junction stones, which were not included in the training set. The AI system surpassed human specialists in speed, analyzing 150 CT images in 13 s, versus 38.6 s for evaluation by urologists and 23 h for formal reading. The AI system calculated stone volume in 0.2 s, versus 77 s for calculation by urologists.
    CONCLUSIONS: Our AI system, which uses advanced deep learning, assists in diagnosing urolithiasis with 94% accuracy in real clinical settings and has potential for rapid diagnosis using standard consumer-grade GPUs.
    RESULTS: We developed a new AI (artificial intelligence) system that can quickly and accurately detect kidney stones in CT (computed tomography) scans. Testing showed that this system is highly effective, with accuracy of 94% for real cases in the emergency department. It is much faster than traditional methods and provides rapid and reliable results to help doctors in making better treatment decisions for their patients.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本研究论文对彩色图像处理技术和深度学习算法在开发专门为8球台球设计的机器人视觉系统中的应用进行了全面的研究。台球运动,随着各种游戏和球的安排,为机器人视觉系统提出了独特的挑战。所提出的方法通过两个主要部分来解决这些挑战:物体检测和球模式识别。最初,采用了一种鲁棒的算法,利用颜色空间变换和阈值技术来检测台球。然后通过战略性的裁剪和隔离主要桌子区域来确定台球桌的位置。关键阶段涉及识别球图案以区分实心和条纹球的复杂任务。为了实现这一点,使用了改进的卷积神经网络,利用Xception网络优化的创新算法称为改进的混沌非洲秃鹰优化(ICAVO)算法。ICAVO算法通过有效地探索解空间和避免局部最优来提高Xception网络的性能。这项研究的结果表明,识别准确性显着提高,Xception/ICAVO模型对实心和条纹球都实现了显着的识别率。这为开发更复杂,更高效的台球机器人铺平了道路。这项研究的意义超出了8球台球,突出了先进的机器人视觉系统在各种应用中的潜力。彩色图像处理的成功集成,深度学习,优化算法验证了所提方法的有效性。这项研究具有深远的意义,不仅仅是台球。尖端的机器人视觉技术可用于检测和跟踪不同领域的物体,改造工业自动化和监控设施。通过结合彩色图像处理,深度学习,和优化算法,证明了该系统的有效性和灵活性。这种创新方法为在各个行业创建先进和高效的机器人视觉系统奠定了基础。
    This research paper presents a comprehensive investigation into the utilization of color image processing technologies and deep learning algorithms in the development of a robot vision system specifically designed for 8-ball billiards. The sport of billiards, with its various games and ball arrangements, presents unique challenges for robotic vision systems. The proposed methodology addresses these challenges through two main components: object detection and ball pattern recognition. Initially, a robust algorithm is employed to detect the billiard balls using color space transformation and thresholding techniques. This is followed by determining the position of the billiard table through strategic cropping and isolation of the primary table area. The crucial phase involves the intricate task of recognizing ball patterns to differentiate between solid and striped balls. To achieve this, a modified convolutional neural network is utilized, leveraging the Xception network optimized by an innovative algorithm known as the Improved Chaos African Vulture Optimization (ICAVO) algorithm. The ICAVO algorithm enhances the Xception network\'s performance by efficiently exploring the solution space and avoiding local optima. The results of this study demonstrate a significant enhancement in recognition accuracy, with the Xception/ICAVO model achieving remarkable recognition rates for both solid and striped balls. This paves the way for the development of more sophisticated and efficient billiards robots. The implications of this research extend beyond 8-ball billiards, highlighting the potential for advanced robotic vision systems in various applications. The successful integration of color image processing, deep learning, and optimization algorithms shows the effectiveness of the proposed methodology. This research has far-reaching implications that go beyond just billiards. The cutting-edge robotic vision technology can be utilized for detecting and tracking objects in different sectors, transforming industrial automation and surveillance setups. By combining color image processing, deep learning, and optimization algorithms, the system proves its effectiveness and flexibility. The innovative approach sets the stage for creating advanced and productive robotic vision systems in various industries.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    水稻是一种重要的主食作物,养活了世界一半以上的人口。优化水稻育种以提高粮食产量对于全球粮食安全至关重要。标题-日期相关或开花时间相关特征,是决定产量潜力的关键因素。然而,这些性状的传统人工表型方法耗时耗力。
    在这里,我们展示了来自无人机(UAV)的航拍图像,当与基于深度学习的穗检测相结合时,实现了与抽穗期相关性状的高通量表型鉴定。我们系统地评估了水稻穗计数的各种最新物体探测器,并将YOLOv8-X确定为最佳探测器。
    将YOLOv8-X应用于294个水稻重组自交系(RIL)的无人机时间序列图像,可以对六个与抽穗日期相关的性状进行准确的定量。利用这些表型,我们确定了数量性状基因座(QTL),包括经过验证的基因座和新基因座,与标题日期相关联。
    我们优化的无人机表型和计算机视觉管道可能有助于对抽穗期相关基因进行可扩展的分子鉴定,并指导水稻产量和适应性的提高。
    UNASSIGNED: Rice (Oryza sativa) serves as a vital staple crop that feeds over half the world\'s population. Optimizing rice breeding for increasing grain yield is critical for global food security. Heading-date-related or Flowering-time-related traits, is a key factor determining yield potential. However, traditional manual phenotyping methods for these traits are time-consuming and labor-intensive.
    UNASSIGNED: Here we show that aerial imagery from unmanned aerial vehicles (UAVs), when combined with deep learning-based panicle detection, enables high-throughput phenotyping of heading-date-related traits. We systematically evaluated various state-of-the-art object detectors on rice panicle counting and identified YOLOv8-X as the optimal detector.
    UNASSIGNED: Applying YOLOv8-X to UAV time-series images of 294 rice recombinant inbred lines (RILs) allowed accurate quantification of six heading-date-related traits. Utilizing these phenotypes, we identified quantitative trait loci (QTL), including verified loci and novel loci, associated with heading date.
    UNASSIGNED: Our optimized UAV phenotyping and computer vision pipeline may facilitate scalable molecular identification of heading-date-related genes and guide enhancements in rice yield and adaptation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    X光成像机广泛应用于边境管制检查站或公共交通,用于行李扫描和检查。深度学习的最新进展使X射线成像结果的自动对象检测能够大大降低人工成本。与自然图像上的任务相比,X射线检查的物体检测通常更具挑战性,由于X射线图像的大小和长宽比不同,小目标对象在冗余背景区域内的随机位置,等。在实践中,我们表明,直接应用现成的基于深度学习的X射线图像检测算法可以是非常耗时和无效的。为此,我们提出了一种任务驱动的裁剪方案,被称为TDC,用于通过X射线图像改进深度图像检测算法,以实现高效和有效的行李检查。而不是处理整个X射线图像进行物体检测,我们提出了两阶段战略,它首先自适应地裁剪X射线图像,只保留与任务相关的区域,即,行李区进行安全检查。特定于任务的深度特征提取器用于快速识别每个X射线图像像素的重要性。仅保留有用且与检测任务相关的区域并将其传递给后续深度检测器。因此,可变比例的X射线图像被缩小到相同的尺寸和纵横比,这可以实现更有效的深度检测管道。此外,为了衡量X射线图像检测算法的有效性,我们提出了一种用于X射线图像检测的新数据集,被称为SIXray-D,基于流行的SIXray数据集。在SIXray-D中,我们提供对象类和边界框的完整和更准确的注释,这使得有监督的X射线检测方法能够进行模型训练。我们的结果表明,我们提出的TDC算法可以有效地提高流行的检测算法,通过实现更好的检测mAP或减少运行时间。
    X-ray imaging machines are widely used in border control checkpoints or public transportation, for luggage scanning and inspection. Recent advances in deep learning enabled automatic object detection of X-ray imaging results to largely reduce labor costs. Compared to tasks on natural images, object detection for X-ray inspection are typically more challenging, due to the varied sizes and aspect ratios of X-ray images, random locations of the small target objects within the redundant background region, etc. In practice, we show that directly applying off-the-shelf deep learning-based detection algorithms for X-ray imagery can be highly time-consuming and ineffective. To this end, we propose a Task-Driven Cropping scheme, dubbed TDC, for improving the deep image detection algorithms towards efficient and effective luggage inspection via X-ray images. Instead of processing the whole X-ray images for object detection, we propose a two-stage strategy, which first adaptively crops X-ray images and only preserves the task-related regions, i.e., the luggage regions for security inspection. A task-specific deep feature extractor is used to rapidly identify the importance of each X-ray image pixel. Only the regions that are useful and related to the detection tasks are kept and passed to the follow-up deep detector. The varied-scale X-ray images are thus reduced to the same size and aspect ratio, which enables a more efficient deep detection pipeline. Besides, to benchmark the effectiveness of X-ray image detection algorithms, we propose a novel dataset for X-ray image detection, dubbed SIXray-D, based on the popular SIXray dataset. In SIXray-D, we provide the complete and more accurate annotations of both object classes and bounding boxes, which enables model training for supervised X-ray detection methods. Our results show that our proposed TDC algorithm can effectively boost popular detection algorithms, by achieving better detection mAPs or reducing the run time.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    语音频率跟随响应(sFFR)越来越多地用于翻译听觉研究。基于统计的自动化sFFR检测可以帮助响应识别,并且在临床和/或研究应用中记录响应时提供停止规则的基础。在这份简短的报告中,sFFR是在安静和语音噪声中从18位正常听力的成人听众中测量的。两种基于统计的自动响应检测方法,F测试和Hotelling的T2(HT2)测试,根据检测精度和测试时间进行了比较。在统计测试和条件下观察到相似的检测准确性,尽管HT2测试时间变化较小。这些发现表明,使用F测试或HT2测试,自动sFFR检测对于在安静和语音噪声中记录的响应是可靠的。未来的研究评估测试性能与不同的刺激和掩蔽是必要的,以确定测试性能的互换性是否延伸到这些条件。
    Speech frequency following responses (sFFRs) are increasingly used in translational auditory research. Statistically-based automated sFFR detection could aid response identification and provide a basis for stopping rules when recording responses in clinical and/or research applications. In this brief report, sFFRs were measured from 18 normal hearing adult listeners in quiet and speech-shaped noise. Two statistically-based automated response detection methods, the F-test and Hotelling\'s T2 (HT2) test, were compared based on detection accuracy and test time. Similar detection accuracy across statistical tests and conditions was observed, although the HT2 test time was less variable. These findings suggest that automated sFFR detection is robust for responses recorded in quiet and speech-shaped noise using either the F-test or HT2 test. Future studies evaluating test performance with different stimuli and maskers are warranted to determine if the interchangeability of test performance extends to these conditions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Objective methods for identifying and quantifying atmospheric blocking have been developed over recent decades, primarily targeting North Atlantic blocks. Differences arise from these methods, leading to changes in the resultant blocking climatology. To understand these differences, and better inform future assessments built on quantitative detection of blocks, this paper examines blocking properties produced by three different objective detection algorithms over the global extratropics. Blocking criteria examined include 500 hPa geopotential height anomaly ( Z ∗ ), column-averaged potential vorticity anomaly ( P V ∗ ), and 500 hPa geopotential height gradient (AGP). Results are analyzed for blocking climatologies and for instantaneous blocking patterns, as well as distributions of block size, speed, duration, and distance traveled. The results emphasize physical characteristics of the flow field and the subsequent blocking regions that emerge; overall, P V ∗ and Z ∗ blocked regions often have higher pattern correlation and spatial similarity, though these two methods also display high agreement with AGP in some instances. Z ∗ finds the largest (and greatest number of) blocked regions, while P V ∗ -detected regions are smallest in all instances except Southern Hemisphere winter. In some cases, P V ∗ tracks a nearby jet streak, leading to differences with height-based algorithms. All three algorithms detect some questionable low-latitude blocks that are stationary and persist but do not impair zonal flow, although at different times. Therefore, careful consideration of the algorithm biases is important in future blocking studies. For example, linking extreme weather to detected blocking could vary substantially depending on the algorithm used.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号