The multi-volume set of LNCS books with volume numbers 15059 up to 15147 constitutes the refereed proceedings of the 18th European Conference on Computer Vision, ECCV 2024, held in Milan, Italy, during September 29-October 4, 2024. The 2387 papers presented in these proceedings were carefully reviewed and selected from a total of 8585 submissions. They deal with topics such as computer vision; machine learning; deep neural networks; reinforcement learning; object recognition; image classification; image processing; object detection; semantic segmentation; human pose estimation; 3d…mehr
The multi-volume set of LNCS books with volume numbers 15059 up to 15147 constitutes the refereed proceedings of the 18th European Conference on Computer Vision, ECCV 2024, held in Milan, Italy, during September 29-October 4, 2024.
The 2387 papers presented in these proceedings were carefully reviewed and selected from a total of 8585 submissions. They deal with topics such as computer vision; machine learning; deep neural networks; reinforcement learning; object recognition; image classification; image processing; object detection; semantic segmentation; human pose estimation; 3d reconstruction; stereo vision; computational photography; neural networks; image coding; image reconstruction; motion estimation.
SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark.- AttnZero: Efficient Attention Discovery for Vision Transformers.- Auto-GAS: Automated Proxy Discovery for Training-free Generative Architecture Search.- Auto-DAS: Automated Proxy Discovery for Training-free Distillation-aware Architecture Search.- UniDream: Unifying Diffusion Priors for Relightable Text-to-3D Generation.- TimeCraft: Navigate Weakly-Supervised Temporal Grounded Video Question Answering via Bi-directional Reasoning.- Spectral Subsurface Scattering for Material Classification.- nuCraft: Crafting High Resolution 3D Semantic Occupancy for Unified 3D Scene Understanding.- Dynamic Neural Radiance Field From Defocused Monocular Video.- PiTe: Pixel-Temporal Alignment for Large Video-Language Model.- CarFormer: Self-Driving with Learned Object-Centric Representations.- FreeDiff: Progressive Frequency Truncation for Image Editing with Diffusion Models.- Plain-Det: A Plain Multi-Dataset Object Detector.- Alternate Diverse Teaching for Semi-supervised Medical Image Segmentation.- Cs2K: Class-specific and Class-shared Knowledge Guidance for Incremental Semantic Segmentation.- Synchronous Diffusion for Unsupervised Smooth Non-Rigid 3D Shape Matching.- Text-Guided Video Masked Autoencoder.- Diffusion Models for Open-Vocabulary Segmentation.- Textual-Visual Logic Challenge: Understanding and Reasoning in Text-to-Image Generation.- EvSign: Sign Language Recognition and Translation with Streaming Events.- QUAR-VLA: Vision-Language-Action Model for Quadruped Robots.- Zero-shot Object Counting with Good Exemplars.- TextDiffuser-2: Unleashing the Power of Language Models for Text Rendering.- SFPNet: Sparse Focal Point Network for Semantic Segmentation on General LiDAR Point Clouds.- PartSTAD: 2D-to-3D Part Segmentation Task Adaptation.- FutureDepth: Learning to Predict the Future Improves Video Depth Estimation.- LLM as Copilot for Coarse-grained Vision-and-Language Navigation.
SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark.- AttnZero: Efficient Attention Discovery for Vision Transformers.- Auto-GAS: Automated Proxy Discovery for Training-free Generative Architecture Search.- Auto-DAS: Automated Proxy Discovery for Training-free Distillation-aware Architecture Search.- UniDream: Unifying Diffusion Priors for Relightable Text-to-3D Generation.- TimeCraft: Navigate Weakly-Supervised Temporal Grounded Video Question Answering via Bi-directional Reasoning.- Spectral Subsurface Scattering for Material Classification.- nuCraft: Crafting High Resolution 3D Semantic Occupancy for Unified 3D Scene Understanding.- Dynamic Neural Radiance Field From Defocused Monocular Video.- PiTe: Pixel-Temporal Alignment for Large Video-Language Model.- CarFormer: Self-Driving with Learned Object-Centric Representations.- FreeDiff: Progressive Frequency Truncation for Image Editing with Diffusion Models.- Plain-Det: A Plain Multi-Dataset Object Detector.- Alternate Diverse Teaching for Semi-supervised Medical Image Segmentation.- Cs2K: Class-specific and Class-shared Knowledge Guidance for Incremental Semantic Segmentation.- Synchronous Diffusion for Unsupervised Smooth Non-Rigid 3D Shape Matching.- Text-Guided Video Masked Autoencoder.- Diffusion Models for Open-Vocabulary Segmentation.- Textual-Visual Logic Challenge: Understanding and Reasoning in Text-to-Image Generation.- EvSign: Sign Language Recognition and Translation with Streaming Events.- QUAR-VLA: Vision-Language-Action Model for Quadruped Robots.- Zero-shot Object Counting with Good Exemplars.- TextDiffuser-2: Unleashing the Power of Language Models for Text Rendering.- SFPNet: Sparse Focal Point Network for Semantic Segmentation on General LiDAR Point Clouds.- PartSTAD: 2D-to-3D Part Segmentation Task Adaptation.- FutureDepth: Learning to Predict the Future Improves Video Depth Estimation.- LLM as Copilot for Coarse-grained Vision-and-Language Navigation.
Es gelten unsere Allgemeinen Geschäftsbedingungen: www.buecher.de/agb
Impressum
www.buecher.de ist ein Internetauftritt der buecher.de internetstores GmbH
Geschäftsführung: Monica Sawhney | Roland Kölbl | Günter Hilger
Sitz der Gesellschaft: Batheyer Straße 115 - 117, 58099 Hagen
Postanschrift: Bürgermeister-Wegele-Str. 12, 86167 Augsburg
Amtsgericht Hagen HRB 13257
Steuernummer: 321/5800/1497